-
Microsoft Research
- China
- https://addf400.github.io/
- @HangboBao
Lists (3)
Sort Name ascending (A-Z)
Stars
Implementation of "Hyperspherical Latents Improve Continuous-Token Autoregressive"
FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory…
Official inference framework for 1-bit LLMs
The leading native Python SSHv2 protocol library.
A PyTorch native platform for training generative AI models
New repo collection for NVIDIA Cosmos: https://github.com/nvidia-cosmos
A Data Streaming Library for Efficient Neural Network Training
Official inference library for pre-processing of Mistral models
Official inference library for Mistral models
[TMLR 2025🔥] A survey for the autoregressive models in vision.
🔥🔥🔥 [IEEE TCSVT] Latest Papers, Codes and Datasets on Vid-LLMs.
A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.
This repo contains the code for 1D tokenizer and generator
SEED-Voken: A Series of Powerful Visual Tokenizers
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
a highly efficient compression algorithm for the n1 implant (neuralink's compression challenge)
Making large AI models cheaper, faster and more accessible
Hackable and optimized Transformers building blocks, supporting a composable construction.