Get the latest tech news
Build your own Siri locally and on-device
That doesn’t spy on you
We’ll use Whisper (tiny model) to transcribe speech, send it through the LLM, parse the response, and call the actual function on-device. This step allows us to deploy the model efficiently on consumer hardware, laptops, phones, even Raspberry PIs (will be a BONUS chapter about this), without needing a GPU at inference time. We’ll write prompts, simulate user requests, auto-verify outputs, and build the data we need to fine-tune the model.
Or read this on Hacker News