They're already getting crushed by the running costs necessary for AI, especially data storage, analysis, and energy. They're trying to get the processors cheaper, since that's the easiest pressure point, hence the recent NVIDIA crash.
Nostr distributes all of those costs. Since your server doesn't have to do all analysis for everyone, but only cater to your own use case and your own data set, and can run asynchronously -- results periodically updating a data file on the server or generating an event for your (private) relay-- it's relatively cheap and you can buy one of the older or less-powerful GPUs for $700-$1500, or just run it on your gaming PC.
We literally just listed the embedding spec in the NIP repo this week, and you've already lost patience. 😅 Like, give us at least a year, okay?
https://github.com/nostr-protocol/nips/blob/master/README.md