-
Huazhong University of Science and Technology
- Wuhan, China
- https://jianyue.tech
Highlights
- Pro
- All languages
- Assembly
- BibTeX Style
- C
- C#
- C++
- CSS
- Cuda
- Dart
- Dockerfile
- Emacs Lisp
- Gnuplot
- Go
- HTML
- Java
- JavaScript
- Jupyter Notebook
- Kotlin
- LLVM
- Lua
- MDX
- MLIR
- Makefile
- Mermaid
- Oz
- Perl
- PostScript
- PowerShell
- Python
- QML
- R
- ReScript
- Ruby
- Rust
- SCSS
- Scala
- Shell
- SystemVerilog
- TLA
- Tcl
- TeX
- TypeScript
- VHDL
- Vala
- Verilog
- Vim Script
- Vue
Starred repositories
🦉 OWL: Optimized Workforce Learning for General Multi-Agent Assistance in Real-World Task Automation
A lightweight, powerful framework for multi-agent workflows
Tongyi Deep Research, the Leading Open-source Deep Research Agent
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Source Han Sans | 思源黑体 | 思源黑體 | 思源黑體 香港 | 源ノ角ゴシック | 본고딕
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
verl: Volcano Engine Reinforcement Learning for LLMs
Train your AI self, amplify you, bridge the world
AKShare is an elegant and simple financial data interface library for Python, built for human beings! 开源财经数据接口库
The official GitHub page for the survey paper "A Survey of Large Language Models".
Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.
A framework for few-shot evaluation of language models.
Self-hosted bookmark manager that is designed be to be minimal, fast, and easy to set up using Docker.
Running large language models on a single GPU for throughput-oriented scenarios.
Matplotlib styles for scientific plotting
The official repo for “Dolphin: Document Image Parsing via Heterogeneous Anchor Prompting”, ACL, 2025.
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Open Source Deep Research Alternative to Reason and Search on Private Data. Written in Python.
Supercharge Your LLM with the Fastest KV Cache Layer
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Benchmarks of approximate nearest neighbor libraries in Python
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
📚A curated list of Awesome LLM/VLM Inference Papers with Codes: Flash-Attention, Paged-Attention, WINT8/4, Parallelism, etc.🎉
g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains
General technology for enabling AI capabilities w/ LLMs and MLLMs