Multi-step planning, code reasoning under uncertainty, decisions that affect the rest of the run. The frontier model in our suite.
Speak.
Command.
Anywhere.
Local Whisper transcription + desktop automation. Zero cloud. Zero latency. System-wide voice control that actually respects your privacy.
You can direct fifteen people in a room. You can't type to them.
Conversations scale. Keyboards do not. The interface a human uses to drive multiple AI agents in parallel is the constraint that decides whether agentic AI is a real tool or a slow demo. Cadence treats voice as the primary input channel for one reason: nothing else keeps up.
- · Serial. One agent at a time.
- · 1.2s per mouse trip (Fitts).
- · Window switch breaks attention.
- · Addressee = where you click.
- ·Parallel. N agents at once.
- ·<200ms first token, on device.
- ·Eyes free. No window switching.
- ·Addressable by name.
What Cadence is, in one paragraph.
Cadence is a voice-first interface for AI agents on macOS. One operator dictates commands by hotkey; Cadence transcribes locally with Whisper, routes the command to one of several Anthropic Claude agents running in parallel, and surfaces results inline. There is no model picker, no cloud transcription, no subscription. The product is built and maintained by Cadence Labs, open-source under the MIT license, and runs exclusively on Claude — Opus, Sonnet, and Haiku.
| Capability | Cadence | Apple Siri | Cloud dictation apps |
|---|---|---|---|
| On-device transcription | Yes (Whisper) | Partial | No |
| Multi-agent orchestration | Yes | No | No |
| System-wide voice command | Yes | Limited | App-scoped |
| Open source | MIT | No | No |
| Powered by | Anthropic Claude | Apple proprietary | Varies |
Spawn. Name. Direct. Approve.
Run multiple Claude agents in parallel — each one named, each one addressable. Voice routes by name; the roster is your mental model. Approvals land inline, in context, never in a separate panel. State is visible at a glance.
One sentence. Three agents updated. Try doing this with a mouse.
How It Works
Loading demo...
Hybrid intelligence meets OS-level control
Optimized Whisper Runtime
Custom-tuned speech recognition optimized for Apple Silicon. Sub-200ms inference with zero network calls.
Adaptive Model Selection
Seamlessly routes between on-device and frontier models. Fast local refinement with optional cloud intelligence.
Screen-Aware Processing
Context engine detects active app, selected text, clipboard state. Knows when you're coding vs. writing vs. messaging.
Universal Text Control
Native macOS APIs for system-wide injection. Terminal, browser, IDE, design tools—no app-specific plugins.
One family. Three models. Calibrated to one voice.
Cadence runs Anthropic Claude exclusively. No GPT. No Gemini. No router. Routing happens automatically, per task type, against one consistent and very good family of models. We made the call once and built the entire product around it.
Day-to-day tool use. Edits, searches, runs commands, summarizes output. The default model behind nearly every action.
Voice intent classification, command routing, fast confidence checks. Latency-critical work where Opus would be overkill.
Get Cadence
Cadence — Apple Silicon Latest
Cadence — Intel Latest
Open Source — Coming Soon
Cadence will be fully open source under the MIT License. Source code, build instructions, and contribution guidelines coming soon. Stay tuned for the public repository launch.