Hedy ships full offline meeting summaries with Qwen 3.5/3.6 on M4 Max
AI meeting app Hedy now runs full transcription, summaries, detailed notes, and chat entirely on-device using llama.cpp and Whisper, with Qwen models from 2B to 35B and no cloud fallback.
Hedy, an AI meeting assistant, shipped full offline meeting processing this week — transcription, summaries, detailed notes, and chat with the meeting all run on-device with Wi-Fi off. The app's founder demonstrated the end-to-end flow on an M4 Max with no network connection, using llama.cpp for inference and Qwen 3.5/3.6 models for language tasks. Speech recognition has always run locally via whisper.cpp and Parakeet; the new release extends that to the entire AI pipeline.
