-
Dadabots
- Boston, MA
- http://dadabots.com
- @dadabots
Stars
Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference
ezkl is an engine for doing inference for deep learning models and other computational graphs in a zk-snark (ZKML). Use it from Python, Javascript, or the command line.
Implementation of HyperDreamBooth: HyperNetworks for Fast Personalization of Text-to-Image Models
State-of-the-art audio codec with 90x compression factor. Supports 44.1kHz, 24kHz, and 16kHz mono/stereo audio.
Download the MusicCaps dataset for music captioning
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable…
open soundstream-ish VAE codecs for downstream neural audio synthesis
Converts an ONNX ML model protobuf from/to text, or tensor from/to text/CSV/raw data. (Windows command line tool)
Table with operator formulas and mappings at a glance, as much ML documentation is missing details
Official implementation of the RAVE model: a Realtime Audio Variational autoEncoder
A directory of what the Metaplex Foundation works on!
Audio bandwidth enhancement with DNNs, addressing filter overfitting
MLSP 2021 - Self-Attention for Audio Super-resolution - Keras implementation
NU-Wave: A Diffusion Probabilistic Model for Neural Audio Upsampling
Official pytorch implementation of the paper: "Catch-A-Waveform: Learning to Generate Audio from a Single Short Example" (NeurIPS 2021)
Implementation of WaveGrad high-fidelity vocoder from Google Brain in PyTorch.
PyTorch implementation of MuseMorphose (published at IEEE/ACM TASLP), a Transformer-based model for music style transfer.
Unofficial PyTorch implementation of Fastformer based on paper "Fastformer: Additive Attention Can Be All You Need"."
Official pytorch implementation of the paper: "SinGAN: Learning a Generative Model from a Single Natural Image"
A convolutional generative audio synthesis model
Source code for models described in the paper "AudioCLIP: Extending CLIP to Image, Text and Audio" (https://arxiv.org/abs/2106.13043)