Context
I journal. I forget to. I wanted to know if a voice-first journaling assistant — something that just listens and prompts gently — could make the practice stick. No login, no cloud, no subscription. A single page, a microphone, a small model.
Conversational AI experiment.
I journal. I forget to. I wanted to know if a voice-first journaling assistant — something that just listens and prompts gently — could make the practice stick. No login, no cloud, no subscription. A single page, a microphone, a small model.
Build a working prototype of a voice-first journaling assistant that runs entirely in the browser; no data leaves the device. The conversation should feel turn-taking, not interrogative.
Whisper.cpp compiled to WASM for transcription. A small GGUF-quantised LLM served from a local llama.cpp instance for the response. Web Audio API for capture, Web Speech API for output. The whole loop sub-second on an M-series Mac.
Working weekend prototype. Used it on myself for a week — the voice-first surface really does change what you talk about. Convinced me there's a real product here, but not as a product I want to ship alone.
The hardest part wasn't the model — it was the silence handling. Voice-first interfaces need a sense of pacing, and 'when does the assistant know I'm done' is unsolved on the open web.