Stars
The repo is finally unlocked. enjoy the party! The fastest repo in history to surpass 100K stars ⭐. Join Discord: https://discord.gg/5TUQKqFWd Built in Rust using oh-my-codex.
TypeScript multi-agent framework — one runTeam() call from goal to result. Auto task decomposition, parallel execution. 3 dependencies, deploys anywhere Node.js runs.
📚 《从零开始构建智能体》——从零开始的智能体原理与实践教程
Provide with pre-build flash-attention package wheels on Linux and Windows platforms using GitHub Actions
每个人都能看懂的大模型知识分享,LLMs春/秋招大模型面试前必看,让你和面试官侃侃而谈
主要记录大语言大模型(LLMs) 算法(应用)工程师相关的知识及面试题
🟣 LLMs interview questions and answers to help you prepare for your next machine learning and data science interview in 2026.
[NeurIPS 2025] Reinforcing Spatial Reasoning in Vision-Language Models with Interwoven Thinking and Visual Drawing
Resources and paper list for "Thinking with Images for LVLMs". This repository accompanies our survey on how LVLMs can leverage visual information for complex reasoning, planning, and generation.
[ICLR 2026] "VTool-R1: VLMs Learn to Think with Images via Reinforcement Learning on Multimodal Tool Use"
This repository contains a curated collection of 300+ case studies from over 80 companies, detailing practical applications and insights into machine learning (ML) system design. The contents are o…
[NeurIPS 2025] ChartMuseum: Testing Visual Reasoning Capabilities of Large Vision-Language Models
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
[EMNLP 2025] Distill Visual Chart Reasoning Ability from LLMs to MLLMs
Awesome-Paper-list: Visualization meets LLM
Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"
An open-source implementaion for fine-tuning Qwen-VL series by Alibaba Cloud.
Synthetic data curation for post-training and structured data extraction
📖 A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, Phi4, ...)…
✨✨Latest Advances on Multimodal Large Language Models
Papers about Hallucination in Multi-Modal Large Language Models (MLLMs)
MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]