-
The University of Texas at Austin
- Austin, TX, USA
-
00:29
(UTC -05:00) - zhenyu.gallery
- @KyriectionZhang
Lists (8)
Sort Name ascending (A-Z)
🦾 Benchmarking
LLM Hospital💎 Efficient ML
Prune & Sparse & Quantization & KD & NAS🤖 General Topics
Architectures & Optimization & BlockChain & SSL & Speech & Recsys💍 Large Language Models
Next Step of LLMs🚀 My Stack
Open-source of Our Works💁 Quantum ML
ML for Quantum & Quantum for ML🗼 Toolbox
Visualization & Coding Tool🚩 Trustworthy ML
OoD & Adversarial & BackdoorStars
MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
NanoGPT (124M) quality in 2.67B tokens
Code for the paper: Why Transformers Need Adam: A Hessian Perspective
A bibliography and survey of the papers surrounding o1
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Retrieval-Augmented Theorem Provers for Lean
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
A new markup-based typesetting system that is powerful and easy to learn.
Easiest way to build custom agents, in a no-code notion style editor, using simple macros.
Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
The official repo for "LLoCo: Learning Long Contexts Offline"
Improving Alignment and Robustness with Circuit Breakers
Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Official implementation of paper "On the Diagram of Thought" (https://arxiv.org/abs/2409.10038)
LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.
High performance AI inference stack. Built for production. @ziglang / @openxla / MLIR / @bazelbuild
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.
METIS - Serial Graph Partitioning and Fill-reducing Matrix Ordering
Main development repository for GAP - Groups, Algorithms, Programming, a System for Computational Discrete Algebra
This repo is based on https://github.com/jiaweizzhao/GaLore
Breaking Throughput-Latency Trade-off for Long Sequences with Speculative Decoding