TeaPoly/CE-OptimizedLoss

Optimized loss based on cross-entropy (CE), like MWER (minimum WER) Loss with beam search and negative sampling strategy, Smoothed Max Pooling Loss.

30
/ 100
Emerging

This project provides advanced techniques to refine how speech recognition models learn, especially when aiming for high accuracy in transcribing spoken language. It takes in the raw output (logits) from a speech model and helps fine-tune it by providing more accurate feedback during the training process. This is for machine learning engineers or researchers building or improving speech-to-text systems.

No commits in the last 6 months.

Use this if you are training speech recognition models and want to optimize their performance to minimize word error rates.

Not ideal if you are looking for a pre-trained speech recognition model or a tool for basic speech transcription.

speech-to-text automatic-speech-recognition model-training natural-language-processing audio-transcription
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 8 / 25
Community 16 / 25

How are scores calculated?

Stars

24

Forks

6

Language

Python

License

Last pushed

Oct 11, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/TeaPoly/CE-OptimizedLoss"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.