-
Tsinghua University
- Beijing
-
07:17
(UTC +08:00) - robertluo1.github.io
Lists (3)
Sort Name ascending (A-Z)
Stars
Turn any PDF or image document into structured data for your AI. A powerful, lightweight OCR toolkit that bridges the gap between images/PDFs and LLMs. Supports 100+ languages.
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
No fortress, purely open ground. OpenManus is Coming.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Open-Sora: Democratizing Efficient Video Production for All
A generative world for general-purpose robotics & embodied AI learning.
Generative Models by Stability AI
Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.
Fully open reproduction of DeepSeek-R1
Official inference repo for FLUX.1 models
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Fast and memory-efficient exact attention
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
verl: Volcano Engine Reinforcement Learning for LLMs
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Janus-Series: Unified Multimodal Understanding and Generation Models
Qwen3-Coder is the code version of Qwen3, the large language model series developed by Qwen team.
Wan: Open and Advanced Large-Scale Video Generative Models
Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.
High-Resolution 3D Assets Generation with Large Scale Hunyuan3D Diffusion Models.
Easy-to-use and powerful LLM and SLM library with awesome model zoo.
Minimal reproduction of DeepSeek R1-Zero
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3, Qwen3-MoE, DeepSeek-R1, GLM4.5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, …
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)