IliaZenkov/sklearn-audio-classification

An in-depth analysis of audio classification on the RAVDESS dataset. Feature engineering, hyperparameter optimization, model evaluation, and cross-validation with a variety of ML techniques and MLP

44
/ 100
Emerging

This project helps developers understand how to classify audio data, specifically for recognizing emotions from speech. It takes raw audio snippets and, through various signal processing and machine learning techniques, outputs a prediction of the emotion conveyed. Data scientists, machine learning engineers, and researchers working with audio will find this useful for learning foundational concepts.

No commits in the last 6 months.

Use this if you are a developer looking for an educational resource to learn the fundamentals of audio feature engineering and classification with classical ML models and MLPs.

Not ideal if you need an out-of-the-box, production-ready solution for state-of-the-art audio classification or complex deep neural networks beyond an MLP.

audio-analysis speech-emotion-recognition machine-learning-education feature-engineering data-science
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 19 / 25

How are scores calculated?

Stars

79

Forks

20

Language

Jupyter Notebook

License

MIT

Last pushed

Nov 05, 2020

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/IliaZenkov/sklearn-audio-classification"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.