@957492b3 @27e3ee9e the OSS LLM world is thriving! And, lots of people are working on getting these working locally without melting the machine :) Check out https://bootcamp.uxdesign.cc/a-complete-guide-to-running-local-llm-models-3225e4913620 I don’t know quite what the term would be other than “local LLM” - however, regarding ML models with very specific purposes running on little boards, look up TinyML. Thriving community there as well.
@63063633 @27e3ee9e I'm familiar with TinyML, it can run on RPis but they are significantly smaller than the industrial ones running in the cloud. I realize there may be no clear ways of measuring this yet but even using model size as a proxy, I've heard that ChatGPT 4 is in the terabytes. Of course, it's not even clear we WANT something as comprehensive at that for many local tasks but I would expect even a reasonable language model is likely to be pretty large
@957492b3 @27e3ee9e yes, my understanding is that the models that run locally on consumer hardware have been trained on specific domains or using compression techniques and thus, smaller. But, tbh, I am on the AI/ML learning curve (I’m a UXer) and ingesting a lot of the the tech and terminology for the first time. Also interesting: https://github.com/KillianLucas/open-interpreter/ https://openinterpreter.com