FabianGroeger96/deep-embedded-music

Creation of an embedding space using unsupervised triplet loss and Tile2Vec that can be used for a variety of downstream tasks

31
/ 100
Emerging

This project helps you organize and understand large collections of audio by grouping similar-sounding tracks together. It takes raw audio files and generates a 'map' where the distance between points represents how similar two sounds are. Audio engineers, music curators, or sound archivists can use this to quickly identify patterns or create playlists based on acoustic similarity.

No commits in the last 6 months.

Use this if you need to automatically categorize or find similarities between audio files without manually tagging them, especially for tasks like music genre classification or environmental sound analysis.

Not ideal if you need a pre-trained, off-the-shelf solution for audio classification without any model training or if your primary need is for human-understandable labels rather than similarity scores.

audio-analysis music-categorization sound-recognition acoustic-similarity media-asset-management
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 9 / 25

How are scores calculated?

Stars

18

Forks

2

Language

Jupyter Notebook

License

MIT

Last pushed

Aug 09, 2021

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/FabianGroeger96/deep-embedded-music"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.