-
Centre for Digital Music, QMUL
-
14:40
(UTC) - @mimbres1
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
🔊 Text-Prompted Generative Audio Model
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable…
A multi-voice TTS system trained with an emphasis on quality
Official inference library for Mistral models
🦙 LaMa Image Inpainting, Resolution-robust Large Mask Inpainting with Fourier Convolutions, WACV 2022
Accepted as [NeurIPS 2024] Spotlight Presentation Paper
A scikit-learn compatible neural network library that wraps PyTorch
Notebooks for my "Deep Learning with TensorFlow 2 and Keras" course
[ICML 2024] Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (RPG)
JonathanFly / bark
Forked from suno-ai/bark🚀 BARK INFINITY GUI CMD 🎶 Powered Up Bark Text-prompted Generative Audio Model
Pix2Seq codebase: multi-tasks with generative modeling (autoregressive and diffusion)
[ICLR 2026] TangoFlux: Super Fast and Faithful Text to Audio Generation with Flow Matching
DiffusionFastForward: a free course and experimental framework for diffusion-based generative models
Code for the paper "LLark: A Multimodal Instruction-Following Language Model for Music" by Josh Gardner, Simon Durand, Daniel Stoller, and Rachel Bittner.
Understanding Training Dynamics of Deep ReLU Networks
Tegridy MIDI Dataset for precise and effective Music AI models creation.
A dataset of 222 digital musical scores aligned with 1068 performances (more than 92 hours) of Western classical piano music.
Steerable discovery of neural audio effects
Official codes and models of the paper "Auffusion: Leveraging the Power of Diffusion and Large Language Models for Text-to-Audio Generation"
Repository for code used in the xVal paper
A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Tra…
Variational Autoencoder in the mel-spectrogram domain for one-shot audio synthesis
Repository for analysis and experiments in the BigCode project.
a notebook containing scripts, documentation, and examples for finetuning musicgen
Memory efficient MAML using gradient checkpointing
Source code for "MusCaps: Generating Captions for Music Audio" (IJCNN 2021)
A Jupyter book accompanying the ISMIR 2023 tutorial Introduction to DIfferentiable Audio Synthesiser Programming