Integrates fully offline text-to-speech and speech-to-text capabilities into coding assistants via the Model Context Protocol (MCP).
Sponsored
VoiceSmith empowers your local AI coding assistants with a natural voice and the ability to listen, all without relying on cloud services. By leveraging Kokoro ONNX for text-to-speech and faster-whisper for speech-to-text, it offers 54 distinct voices and robust voice activity detection, ensuring a private and responsive conversational experience. It seamlessly integrates with IDEs like Claude Code, Cursor, and Codex, allowing for multi-session voice interactions and custom voice selections, enhancing productivity and the overall coding assistant experience.
Key Features
01Local speech-to-text using faster-whisper
0254 distinct voices via Kokoro ONNX (fully offline TTS)
034 GitHub stars
04Seamless integration with Claude Code, Cursor, and Codex
05Voice activity detection with Silero VAD
06Multi-session voice support for Claude Code
Use Cases
01Providing spoken summaries and questions from AI assistants during coding
02Allowing voice commands and responses to interact with AI development tools
03Enabling conversational interfaces for AI coding assistants