OpenGVLab/LORIS
[ICML2023] Long-Term Rhythmic Video Soundtracker
This tool generates synchronized soundtracks for long rhythmic videos like dance performances, floor exercises, or figure skating. You provide a video of a rhythmic activity, and it outputs an audio waveform that matches the visual cues and rhythm. It's designed for creators or analysts who need to add realistic, context-aware sound to their visual content.
No commits in the last 6 months.
Use this if you need to automatically generate long, rhythmically aligned soundtracks for videos of performances like dance or sports routines.
Not ideal if you need to generate sound for non-rhythmic videos, or if you require fine-grained manual control over musical composition.
Stars
62
Forks
1
Language
Python
License
MIT
Category
Last pushed
Jul 28, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/OpenGVLab/LORIS"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
PrunaAI/pruna
Pruna is a model optimization framework built for developers, enabling you to deliver faster,...
bytedance/LatentSync
Taming Stable Diffusion for Lip Sync!
haoheliu/AudioLDM-training-finetuning
AudioLDM training, finetuning, evaluation and inference.
Text-to-Audio/Make-An-Audio
PyTorch Implementation of Make-An-Audio (ICML'23) with a Text-to-Audio Generative Model
teticio/audio-diffusion
Apply diffusion models using the new Hugging Face diffusers package to synthesize music instead...