alphacep/vosk
VOSK Speech Recognition Toolkit
Vosk helps you continuously improve your speech recognition models by leveraging a vast database of audio fingerprints. You input audio files and their corresponding phonetic transcriptions, and the system learns from this data. The output is a more accurate speech recognition model that adapts over time. This is for researchers, engineers, or linguists building highly specialized and evolving speech recognition systems.
493 stars. Used by 6 other packages. No commits in the last 6 months. Available on PyPI.
Use this if you need to train a speech recognition system on extremely large datasets quickly and want the ability to easily correct recognition errors by adding more samples.
Not ideal if you expect your model to fit into the memory of a single server or need strong generalization capabilities for entirely new, unseen acoustic conditions without additional training data.
Stars
493
Forks
56
Language
C
License
Apache-2.0
Category
Last pushed
Jul 13, 2022
Commits (30d)
0
Dependencies
5
Reverse dependents
6
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/alphacep/vosk"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
k2-fsa/sherpa-onnx
Speech-to-text, text-to-speech, speaker diarization, speech enhancement, source separation, and...
ccoreilly/vosk-browser
A speech recognition library running in the browser thanks to a WebAssembly build of Vosk
alphacep/vosk-server
WebSocket, gRPC and WebRTC speech recognition server based on Vosk and Kaldi libraries
solyarisoftware/voskJs
Vosk ASR offline engine API for NodeJs developers. With a simple HTTP ASR server.
alphacep/vosk-asterisk
Speech Recognition in Asterisk with Vosk Server