-
University of Toronto
- Toronto
-
12:16
(UTC -05:00) - https://rexxxx1234.github.io/
- @RexMa9
- in/rex-ma-20a455113
Stars
A high-throughput and memory-efficient inference and serving engine for LLMs
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
Official Code for DragGAN (SIGGRAPH 2023)
Fully open reproduction of DeepSeek-R1
Graph Neural Network Library for PyTorch
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Janus-Series: Unified Multimodal Understanding and Generation Models
Convert Machine Learning Code Between Frameworks
Python package built to ease deep learning on graph, on top of existing DL frameworks.
Open source code for AlphaFold 2.
Official code for "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
Minimal reproduction of DeepSeek R1-Zero
A framework for few-shot evaluation of language models.
Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)
ImageBind One Embedding Space to Bind Them All
A Collection of Variational Autoencoders (VAE) in PyTorch.
Large World Model -- Modeling Text and Video with Millions Context
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Adding guardrails to large language models.