Lists (1)
Sort Name ascending (A-Z)
Starred repositories
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
Fast and memory-efficient exact attention
Build Real-Time Knowledge Graphs for AI Agents
SGLang is a fast serving framework for large language models and vision language models.
Use ChatGPT to summarize the arXiv papers. 全流程加速科研,利用chatgpt进行论文全文总结+专业翻译+润色+审稿+审稿回复
Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
A high-performance, zero-overhead, extensible Python compiler with built-in NumPy support
FauxPilot - an open-source alternative to GitHub Copilot server
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
Python Implementation of Reinforcement Learning: An Introduction
Ongoing research training transformer models at scale
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Minimal reproduction of DeepSeek R1-Zero
An open-source tool-augmented conversational language model from Fudan University
The official GitHub page for the survey paper "A Survey of Large Language Models".
Retrieval and Retrieval-augmented LLMs
Large Language Model Text Generation Inference
Implementation of Denoising Diffusion Probabilistic Model in Pytorch
A Python Library for Outlier and Anomaly Detection, Integrating Classical and Deep Learning Techniques
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
Running large language models on a single GPU for throughput-oriented scenarios.
Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 17+ clouds, or on-prem).
An Easy-to-use, Scalable and High-performance RLHF Framework based on Ray (PPO & GRPO & REINFORCE++ & vLLM & Ray & Dynamic Sampling & Async Agentic RL)
Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
A faster pytorch implementation of faster r-cnn
🚴 Call stack profiler for Python. Shows you why your code is slow!