-
XDU
-
15:45
(UTC +08:00) - https://charent.github.io/
Stars
📚 Freely available programming books
🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
🦜🔗 The platform for reliable agents.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
1 min voice data can also be used to train a good TTS model! (few shot voice cloning)
Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and…
🤗 The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation tools
SGLang is a fast serving framework for large language models and vision language models.
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
🤖 可 DIY 的 多模态 AI 聊天机器人 | 🚀 快速接入 微信、 QQ、Telegram、等聊天平台 | 🦈支持DeepSeek、Grok、Claude、Ollama、Gemini、OpenAI | 工作流系统、网页搜索、AI画图、人设调教、虚拟女仆、语音对话 |
Train transformer language models with reinforcement learning.
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
The official GitHub page for the survey paper "A Survey of Large Language Models".
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discr…
CodeGeeX2: A More Powerful Multilingual Code Generation Model
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
LLM training code for Databricks foundation models
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
🤗 Evaluate: A library for easily evaluating machine learning models and datasets.
PaddleOCR inference in PyTorch. Converted from [PaddleOCR](https://github.com/PaddlePaddle/PaddleOCR)