self-supervised-speech-recognition and wav2vec2-huggingface-demo

These are ecosystem siblings—one is a pretrained model checkpoint while the other is a demonstration application showing how to use wav2vec2 models via Hugging Face's Transformers library for inference.

Maintenance 0/25
Adoption 10/25
Maturity 8/25
Community 24/25
Maintenance 0/25
Adoption 7/25
Maturity 16/25
Community 18/25
Stars: 379
Forks: 116
Downloads:
Commits (30d): 0
Language: Python
License:
Stars: 29
Forks: 14
Downloads:
Commits (30d): 0
Language: Jupyter Notebook
License: Apache-2.0
No License Stale 6m No Package No Dependents
Stale 6m No Package No Dependents

About self-supervised-speech-recognition

mailong25/self-supervised-speech-recognition

speech to text with self-supervised learning based on wav2vec 2.0 framework

This helps speech recognition practitioners develop highly accurate speech-to-text models for any language, even with limited transcribed audio. You provide audio files (both labeled with transcripts and unlabeled) and text data, and it outputs a custom speech recognition model. This is for data scientists, linguists, or AI engineers who need to build robust transcription systems for specific languages or domains.

speech-to-text voice-recognition natural-language-processing audio-transcription custom-language-models

About wav2vec2-huggingface-demo

bhattbhavesh91/wav2vec2-huggingface-demo

Speech to Text with self-supervised learning based on wav2vec 2.0 framework using Hugging Face's Transformer

This project helps convert spoken audio into written text, making it easier to analyze recordings or create transcripts. You feed it an audio file, and it outputs the corresponding text. This is useful for anyone who needs to quickly transcribe speeches, interviews, or voice notes.

audio-transcription voice-to-text meeting-notes interview-analysis content-creation

Scores updated daily from GitHub, PyPI, and npm data. How scores work