yufan-aslp/AliMeeting

The project is associated with the recently-launched ICASSP 2022 Multi-channel Multi-party Meeting Transcription Challenge (M2MeT) to provide participants with baseline systems for speech recognition and speaker diarization in conference scenario.

34
/ 100
Emerging

This project offers baseline systems for transcribing multi-speaker, multi-channel meeting audio. It takes raw meeting recordings and outputs two key pieces of information: a detailed transcription of who said what and when (speaker diarization results in RTTM files) and the actual text of the speech (ASR results measured by Character Error Rate). This is ideal for researchers and engineers working on improving speech recognition and speaker separation in complex meeting environments.

135 stars. No commits in the last 6 months.

Use this if you are a speech researcher or engineer aiming to develop or benchmark advanced systems for transcribing spoken content from multi-participant conference calls or meetings.

Not ideal if you are looking for an out-of-the-box application to transcribe your meetings without significant technical expertise or further development.

speech-recognition speaker-diarization meeting-transcription audio-processing natural-language-processing
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 16 / 25

How are scores calculated?

Stars

135

Forks

18

Language

Python

License

Last pushed

Jun 10, 2022

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/yufan-aslp/AliMeeting"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.