Stars
High quality text-to-speech based on StyleTTS 2.
Community-driven advanced graphics modifications for AE, SE and VR.
HiFTNet: A Fast High-Quality Neural Vocoder with Harmonic-plus-Noise Filter and Inverse Short Time Fourier Transform
Some words that LLM regularly uses
Inference and training library for high-quality TTS models.
An SKSE plugin dll to address a bug in papyrus VM garbage collector that cause game to only cleanup single array item per VM frame
ELLA: Equip Diffusion Models with LLM for Enhanced Semantic Alignment
State-of-the-art audio codec with 90x compression factor. Supports 44.1kHz, 24kHz, and 16kHz mono/stereo audio.
[ICLR 2026] When it comes to optimizers, it's always better to be safe than sorry
JPEG Artifact Correction using Denoising Diffusion Restoration Models -- Official Code Repository
Textual Inversion for Stable Diffusion XL 1.0
Official repository of STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech, INTERSPEECH 2021
PyTorch implementation of A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text (EMNLP 2019)
This repository contains some of the latest data augmentation techniques and optimizers for image classification using pytorch and the CIFAR10 dataset
Zero-Shot Speech Editing and Text-to-Speech in the Wild
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
The Prodigy optimizer and its variants for training neural networks.
AuraSR: GAN-based Super-Resolution for real-world
Lumina-T2X is a unified framework for Text to Any Modality Generation
Paper and code for Gradient Descent: The Ultimate Optimizer
Implementation of Voicebox, new SOTA Text-to-speech network from MetaAI, in Pytorch
ResShift: Efficient Diffusion Model for Image Super-resolution by Residual Shifting (NeurIPS@2023 Spotlight, TPAMI@2024)
📋 A list of open LLMs available for commercial use.
Official repository of https://doi.org/10.1109/TASLP.2022.3167258. More up-to-date code is in "refactor" branch.