archinetai/audio-diffusion-pytorch
Audio generation using diffusion models, in PyTorch.
This is a comprehensive toolkit for anyone working with audio generation using advanced AI models. You can create new audio from scratch, generate audio based on text descriptions, or enhance existing low-quality audio. It takes in audio waveforms or text prompts and outputs high-quality, synthesized audio, making it useful for sound designers, music producers, or researchers in audio AI.
2,094 stars. Used by 1 other package. No commits in the last 6 months. Available on PyPI.
Use this if you are a machine learning researcher or developer focusing on creating, enhancing, or transforming audio using diffusion models and need a flexible, PyTorch-based library.
Not ideal if you are looking for pre-trained models or a ready-to-use application for immediate audio generation without any development work.
Stars
2,094
Forks
178
Language
Python
License
MIT
Category
Last pushed
Jun 12, 2023
Commits (30d)
0
Dependencies
6
Reverse dependents
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/archinetai/audio-diffusion-pytorch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
huggingface/diffusers
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.
bghira/SimpleTuner
A general fine-tuning kit geared toward image/video/audio diffusion models.
mcmonkeyprojects/SwarmUI
SwarmUI (formerly StableSwarmUI), A Modular Stable Diffusion Web-User-Interface, with an...
nateraw/stable-diffusion-videos
Create 🔥 videos with Stable Diffusion by exploring the latent space and morphing between text prompts
TheDesignFounder/DreamLayer
Benchmark diffusion models faster. Automate evals, seeds, and metrics for reproducible results.