- Los Angeles, CA
- All languages
- APL
- Batchfile
- C
- C#
- C++
- CSS
- Chapel
- Clojure
- CoffeeScript
- Common Lisp
- Dart
- Emacs Lisp
- F#
- Fortran
- Frege
- Go
- HCL
- HTML
- Haskell
- Idris
- Java
- JavaScript
- Julia
- Jupyter Notebook
- Kotlin
- Lean
- Lua
- MATLAB
- Mojo
- Nim
- OCaml
- Objective-C
- Objective-C++
- Perl
- PlantUML
- PowerShell
- Python
- R
- RMarkdown
- Racket
- Rich Text Format
- Ruby
- Rust
- Scala
- Scheme
- Shell
- Swift
- TSQL
- TeX
- TypeScript
- Vim Script
- q
Starred repositories
Research code artifacts for Code World Model (CWM) including inference tools, reproducibility, and documentation.
GPU programming related news and material links
A Flexible Framework for Experiencing Heterogeneous LLM Inference/Fine-tune Optimizations
On-device AI across mobile, embedded and edge for PyTorch
Minimalistic 4D-parallelism distributed training framework for education purpose
Minimalistic large language model 3D-parallelism training
RapidFire AI: Rapid AI Customization from RAG to Fine-Tuning
Scalable Python DS & ML, in an API compatible & lightning fast way.
Python actor framework for heterogeneous computing.
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-re…
🚀 Efficient implementations of state-of-the-art linear attention models
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. 🍊YC W23
Turn any computer or edge device into a command center for your computer vision projects.
We write your reusable computer vision tools. 💜
🌷 Run code formatter on buffer contents without moving point, using RCS patches and dynamic programming.
A Rust HTTP server for Python applications
Minimal reproduction of DeepSeek R1-Zero
Tantivy is a full-text search engine library inspired by Apache Lucene and written in Rust
SeekStorm - sub-millisecond full-text search library & multi-tenancy server in Rust
OpenMMLab Foundational Library for Training Deep Learning Models
The world's fastest open query engine for sub-second analytics both on and off the data lakehouse. With the flexibility to support nearly any scenario, StarRocks provides best-in-class performance …