Deepest-Project/MelNet
Implementation of "MelNet: A Generative Model for Audio in the Frequency Domain"
This project helps audio engineers and researchers generate realistic audio by transforming frequency-domain representations (mel spectrograms) into audible sound. You provide training data, which can be raw audio datasets, and it produces new, synthesized audio. It's designed for someone working with speech synthesis or exploring novel sound generation.
210 stars. No commits in the last 6 months.
Use this if you need to generate high-quality, synthetic audio from scratch or from text prompts, particularly for research in speech synthesis or creative sound design.
Not ideal if you need to extend or complete existing audio (primed generation) or if you are looking for a simple, out-of-the-box solution without deep customization.
Stars
210
Forks
41
Language
Python
License
MIT
Category
Last pushed
Jul 25, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/voice-ai/Deepest-Project/MelNet"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
kan-bayashi/ParallelWaveGAN
Unofficial Parallel WaveGAN (+ MelGAN & Multi-band MelGAN & HiFi-GAN & StyleMelGAN) with Pytorch
fatchord/WaveRNN
WaveRNN Vocoder + TTS
shangeth/wavencoder
WavEncoder is a Python library for encoding audio signals, transforms for audio augmentation,...
rishikksh20/iSTFTNet-pytorch
iSTFTNet : Fast and Lightweight Mel-spectrogram Vocoder Incorporating Inverse Short-time Fourier...
seungwonpark/melgan
MelGAN vocoder (compatible with NVIDIA/tacotron2)