Live preview uses a dedicated lightweight model that runs independently from the main transcription engine. This means you get instant visual feedback without sacrificing final accuracy.
A dedicated Parakeet End-of-Utterance model runs on the Apple Neural Engine, processing 320ms audio windows. Produces word-level partial transcripts with ~300ms latency.
When the EOU model is not available, live preview falls back to the main transcription backend processing 2-second chunks during recording.
The EOU engine processes 320ms audio windows on the Neural Engine, producing word-level partial transcripts with near-instant feedback.
EOU runs on the Neural Engine while the main transcription engine uses Metal GPU — no resource contention between preview and final transcription.
Text appears progressively in the UI with a natural typewriter effect, creating a fluid 'words appearing' experience as you speak.
Dictionary corrections are shown with color gradient highlighting in the live preview. Click any highlighted word to see the original.
The EOU engine processes 320ms audio windows on the Neural Engine, producing word-level partial transcripts with approximately 300ms latency — faster than any cloud service.
No. Live preview uses a separate lightweight model. The main transcription engine runs independently and delivers the final refined result when you release the record button.
Yes. Live preview is toggleable in Settings. Disabling it reduces CPU/Neural Engine usage during recording while keeping the main transcription engine active.
The primary EOU preview runs on the Neural Engine (Apple Silicon). When unavailable, Whisperer falls back to the StreamingTranscriber using the main engine with 2-second chunks.