Oddbean new post about | logout
 "Micro$oft BitNet a blazing-fast 1-bit LLM inference framework that runs directly on CPUs.

You can now run 100B parameter models on local devices with up to 6x speed improvements and 82% less energy consumption—all without a GPU!"

https://github.com/microsoft/BitNet 
 I wonder if you can plug it into llama on @umbrel or @Start9 on a system with some GPUs and use it to train data. I have not looked at the llama or the implementation of it on these systems, but if it is based on llama and faster, I bet there is a good use case for optionally running it. Do these installs allow training or are they just running models? 
 Looking forward to see more improvements. I have a Libre Computer Alta which has got a 5 Tops NPU. Imagine having a good LLM on such a small device! 
 I assume the 1-bit models are retarded?