alphacep/vosk

VOSK Speech Recognition Toolkit

58
/ 100
Established

Vosk helps you continuously improve your speech recognition models by leveraging a vast database of audio fingerprints. You input audio files and their corresponding phonetic transcriptions, and the system learns from this data. The output is a more accurate speech recognition model that adapts over time. This is for researchers, engineers, or linguists building highly specialized and evolving speech recognition systems.

493 stars. Used by 6 other packages. No commits in the last 6 months. Available on PyPI.

Use this if you need to train a speech recognition system on extremely large datasets quickly and want the ability to easily correct recognition errors by adding more samples.

Not ideal if you expect your model to fit into the memory of a single server or need strong generalization capabilities for entirely new, unseen acoustic conditions without additional training data.

speech-recognition linguistics audio-processing machine-learning-engineering lifelong-learning
Stale 6m
Maintenance 0 / 25
Adoption 15 / 25
Maturity 25 / 25
Community 18 / 25

How are scores calculated?

Stars

493

Forks

56

Language

C

License

Apache-2.0

Last pushed

Jul 13, 2022

Commits (30d)

0

Dependencies

5

Reverse dependents

6

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/alphacep/vosk"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.