-
@thu-ml, Tsinghua University
- Beijing, China
-
02:11
(UTC +09:00) - https://bingrui-li.github.io/
- @bingruili_
- @bingruil.bsky.social
Stars
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
A high-throughput and memory-efficient inference and serving engine for LLMs
No fortress, purely open ground. OpenManus is Coming.
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
Making large AI models cheaper, faster and more accessible
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
A toolkit for developing and comparing reinforcement learning algorithms.
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (V…
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
Code and documentation to train Stanford's Alpaca models, and generate the data.
A generative world for general-purpose robotics & embodied AI learning.
Fully open reproduction of DeepSeek-R1
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Code for the paper "Language Models are Unsupervised Multitask Learners"
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
DeepSeek Coder: Let the Code Write Itself
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Use ChatGPT to summarize the arXiv papers. 全流程加速科研,利用chatgpt进行论文全文总结+专业翻译+润色+审稿+审稿回复
DeerFlow is a community-driven Deep Research framework, combining language models with tools like web search, crawling, and Python execution, while contributing back to the open-source community.
Tongyi Deep Research, the Leading Open-source Deep Research Agent