kaistmm/Audio-Mamba-AuM
Official Implementation of the work "Audio Mamba: Bidirectional State Space Model for Audio Representation Learning"
Audio-Mamba (AuM) helps you categorize different sounds or spoken words, like identifying specific events in recordings or classifying speech commands. It takes raw audio data as input and produces classifications, telling you what kind of sound or speech is present. Researchers and practitioners working with large audio datasets for sound event detection or speech recognition would find this useful.
167 stars. No commits in the last 6 months.
Use this if you need to classify audio efficiently for tasks like environmental sound monitoring, voice assistant command recognition, or cataloging audio events in large datasets.
Not ideal if your primary goal is audio generation, music composition, or highly nuanced speech-to-text transcription rather than classification.
Stars
167
Forks
20
Language
Python
License
BSD-3-Clause
Category
Last pushed
Nov 24, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/kaistmm/Audio-Mamba-AuM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
kyegomez/VisionMamba
Implementation of Vision Mamba from the paper: "Vision Mamba: Efficient Visual Representation...
SiavashShams/ssamba
[SLT'24] The official implementation of SSAMBA: Self-Supervised Audio Representation Learning...
kkakkkka/MambaTalk
[NeurlPS-2024] The official code of MambaTalk: Efficient Holistic Gesture Synthesis with...
FarnoushRJ/MambaLRP
[NeurIPS 2024] Official implementation of the paper "MambaLRP: Explaining Selective State Space...
niyiyu/DAS-reconstruction
Wavefield Reconstruction of Distributed Acoustic Sensing with Machine Learning