kaistmm/Audio-Mamba-AuM

Official Implementation of the work "Audio Mamba: Bidirectional State Space Model for Audio Representation Learning"

41
/ 100
Emerging

Audio-Mamba (AuM) helps you categorize different sounds or spoken words, like identifying specific events in recordings or classifying speech commands. It takes raw audio data as input and produces classifications, telling you what kind of sound or speech is present. Researchers and practitioners working with large audio datasets for sound event detection or speech recognition would find this useful.

167 stars. No commits in the last 6 months.

Use this if you need to classify audio efficiently for tasks like environmental sound monitoring, voice assistant command recognition, or cataloging audio events in large datasets.

Not ideal if your primary goal is audio generation, music composition, or highly nuanced speech-to-text transcription rather than classification.

audio-classification sound-event-detection speech-recognition acoustic-analysis audio-analytics
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

167

Forks

20

Language

Python

License

BSD-3-Clause

Last pushed

Nov 24, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kaistmm/Audio-Mamba-AuM"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.