micro-musicgen-jungle
Curated and trained by Aaron Abebe.
WARNING: These models WILL sound bad to a lot of people. The goal is not create pleasant sounding music, but to spark creativity by using the weird sounds of Neural Codecs for music production and sampling!
Micro-Musicgen is a new family of super small music generation models focussing on experimental music and latent space exploration capabilities. These models have unique abilities and drawbacks which should enhance creativity when working with them while creating music.
- only unconditional generation: Trained without text-conditioning to reduce model size.
- very fast generation times: ~8secs for 10x 10sec samples.
- permissive licensing: The models are trained from scratch using royalty-free samples and handmade chops, which allows them to be released via the MIT License.
The first entry in the series is a model trained on chopped jungle drum breaks called micro-musicgen-jungle
.
If you find this model interesting, please consider:
Samples
All samples are from a single run, without cherry picking.
Usage
Install my audiocraft fork:
pip install -U git+https://github.com/aaronabebe/audiocraft#egg=audiocraft
Then, you should be able to load this model just like any other musicgen checkpoint here on the Hub:
import torchaudio
from audiocraft.models import MusicGen
from audiocraft.data.audio import audio_write
model = MusicGen.get_pretrained('pharoAIsanders420/micro-musicgen-jungle')
model.set_generation_params(duration=10)
wav = model.generate_unconditional(10)
for idx, one_wav in enumerate(wav):
# Will save under {idx}.wav, with loudness normalization at -14 db LUFS.
audio_write(f'{idx}', one_wav.cpu(), model.sample_rate, strategy="loudness", loudness_compressor=True)
- Downloads last month
- 9