Lists (32)
Sort Name ascending (A-Z)
AI框架与一些工具项目
c++项目
CNN网络
CUDA相关项目
NLP项目
OCR相关项目
opencv相关项目
pdf解析器
tensorrt项目
transformer相关项目
关键点检测项目
动手学深度学习
图像分割项目
图像分类相关项目
图像检索
多模态
对比学习
数据结构与算法刷题
数据集
文本数据合成
文本检测项目
文本识别项目
文档矫正
杂项
深度学习论文精度
目标检测相关项目
直线检测相关仓库
自编码模型
表格识别项目
视频分类
视频切割
骨干网络
Stars
Fast and customizable text tokenization library with BPE and SentencePiece support
A compact implementation of SGLang, designed to demystify the complexities of modern LLM serving systems.
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, Phi4, ...)…
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
UniFace: A Unified Face Analysis Library in Python built on ONNX Runtime
MobileGaze: Real-Time Gaze Estimation models using ResNet 18/34/50, MobileNet v2 and MobileOne s0-s4 | In PyTorch >> ONNX Runtime Inference
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
Analyze the inference of Large Language Models (LLMs). Analyze aspects like computation, storage, transmission, and hardware roofline model in a user-friendly interface.
A tool to modify ONNX models in a visualization fashion, based on Netron and Flask.
🤖FFPA: Extend FlashAttention-2 with Split-D, ~O(1) SRAM complexity for large headdim, 1.8x~3x↑🎉 vs SDPA EA.
cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it
Implement custom operators in PyTorch with cuda/c++
Fast, Flexible and Portable Structured Generation
Community maintained hardware plugin for vLLM on Ascend
校招、秋招、春招、实习好项目,带你从零动手实现支持LLama2/3和Qwen2.5的大模型推理框架。
🤗 Optimum ONNX: Export your model to ONNX and run inference with ONNX Runtime
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
《动手学大模型Dive into LLMs》系列编程实践教程
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling