Oddbean new post about | logout
 If you have a PC, you can run all the cutting-edge language models locally with just one click in your browser. 

Tiny models also work on Android Chrome, and support for iOS will be added, at least for new iPhones.
Temporary demo: https://iefan-ws.web.app/
https://image.nostr.build/ca976421700bb3591eda64d92667fbd108888b875122f23fb333fa81b6897ecf.jpg#m=image%2Fjpeg&alt=Verifiable+file+url&x=443ce0688c16f2c26bd753ba6c16a4e4e2de207427bd16d9ca746206ed1ab00f&size=71841&dim=980x1920&blurhash=%5D55%7Dv%7DsqNFV%5BM%7B%5E%2CtjMyX5M%7B_Nr_S0V%5BRjxbo%7BVubYWBtQepkUaMkBD%25X4sEbZofInVutQaMt7DiX4t8bYofInV%5BtRWBof&ox=ca976421700bb3591eda64d92667fbd108888b875122f23fb333fa81b6897ecf 
 Wow it’s working great! 
 It utilizes the new WebGPU API, harnessing your actual hardware. Apple is also implementing it very soon. It's poised to be a game-changer, especially for game development. 
 I have something even more fascinating, it's still underdeveloped but it definitely has a potential.
nostr:nevent1qqsv0006qurr3wg84zgvqt3h50t0ehlyxm8vk0fe59j7efkphpwpkygpz4mhxue69uhhyetvv9ujumt0wd68ytnsw43qygxx7urh795e65x0j25k22l7hlavqh7xss4eacu3pzwetxu26h2gl5psgqqqqqqsts3h4e 
 In-device models can take care of private events. But otherwise, I am not sure they are that interesting for public events. Constantly updating the local model to match the newest events coming in online is challenging. 
 I was attempting that, and it's quite challenging. So, I will be working on it with Gemini. 

This will focus on private tasks like TTS, PDF summarizer, Basic Q/A etc., using small, task-specific language models with WebGPU. 
 Would be nice if we can set the address of an OpenAI API compatible LLM to give users choice