Stars
An Open Phone Agent Model & Framework. Unlocking the AI Phone for Everyone
A curated list of cutting-edge research papers and resources on Long Chain-of-Thought (CoT) Reasoning with Tools.
Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793
Code for Paper (Preserving Diversity in Supervised Fine-tuning of Large Language Models)
Recent research papers about Foundation Models for Combinatorial Optimization
The LLMOPT project offers a comprehensive set of resources, including the model, dataset, training framework, and inference code, enabling users to fully utilize LLMOPT.
A simple and well styled PPO implementation. Based on my Medium series: https://medium.com/@eyyu/coding-ppo-from-scratch-with-pytorch-part-1-4-613dfc1b14c8.
[ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data generation pipeline!
ORLM: Training Large Language Models for Optimization Modeling
Multilingual Medicine: Model, Dataset, Benchmark, Code
Retrieval-Augmented Theorem Provers for Lean
Implementation of Nougat Neural Optical Understanding for Academic Documents
This dataset code generates mathematical question and answer pairs, from a range of question types at roughly school-level difficulty.
This is a PyTorch reimplementation of Influence Functions from the ICML2017 best paper: Understanding Black-box Predictions via Influence Functions by Pang Wei Koh and Percy Liang.
Everything you need about Active Learning (AL).
Tool for data extraction and interacting with Lean programmatically.
Fast Inference Solutions for BLOOM
deepspeedai / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
800,000 step-level correctness labels on LLM solutions to MATH problems