Whisper-Finetune and whisper-finetune

These are competitors offering overlapping fine-tuning solutions for Whisper ASR, with A differentiating through timestamp-flexible training modes and Web deployment acceleration while B focuses on standard fine-tuning and evaluation workflows.

Whisper-Finetune
56
Established
whisper-finetune
49
Emerging
Maintenance 6/25
Adoption 10/25
Maturity 16/25
Community 24/25
Maintenance 0/25
Adoption 10/25
Maturity 16/25
Community 23/25
Stars: 1,200
Forks: 213
Downloads:
Commits (30d): 0
Language: C
License: Apache-2.0
Stars: 361
Forks: 87
Downloads:
Commits (30d): 0
Language: Python
License: MIT
No Package No Dependents
Stale 6m No Package No Dependents

About Whisper-Finetune

yeyupiaoling/Whisper-Finetune

Fine-tune the Whisper speech recognition model to support training without timestamp data, training with timestamp data, and training without speech data. Accelerate inference and support Web deployment, Windows desktop deployment, and Android deployment

This project helps you improve the accuracy and speed of transcribing audio into text using the Whisper speech recognition system. It allows you to customize the system with your own audio recordings and their corresponding text, even if your data doesn't include exact timing information. The enhanced system can then quickly convert new audio files into accurate written transcripts, and can be deployed in web applications, desktop programs, or Android devices. This is for professionals like journalists, researchers, or content creators who need highly accurate and fast audio transcription tailored to specific languages or accents.

speech-to-text audio-transcription voice-recognition language-processing content-creation

About whisper-finetune

vasistalodagala/whisper-finetune

Fine-tune and evaluate Whisper models for Automatic Speech Recognition (ASR) on custom datasets or datasets from huggingface.

This project helps machine learning engineers and researchers improve Automatic Speech Recognition (ASR) performance for specific languages or accents. It takes audio recordings and their human-transcribed text as input, then customizes an existing Whisper ASR model. The output is a specialized ASR model that is more accurate for your unique audio data.

speech-to-text voice-recognition audio-transcription natural-language-processing machine-learning-engineering

Scores updated daily from GitHub, PyPI, and npm data. How scores work