Stars
pytorch handbook是一本开源的书籍,目标是帮助那些希望和使用PyTorch进行深度学习开发和研究的朋友快速入门,其中包含的Pytorch教程全部通过测试保证可以成功运行
Instruct-tune LLaMA on consumer hardware
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Code for Machine Learning for Algorithmic Trading, 2nd edition.
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
Dopamine is a research framework for fast prototyping of reinforcement learning algorithms.
Gathers machine learning and deep learning models for Stock forecasting including trading bots and simulations
My continuously updated Machine Learning, Probabilistic Models and Deep Learning notes and demos (2000+ slides) 我不间断更新的机器学习,概率模型和深度学习的讲义(2000+页)和视频链接
MiniCPM4 & MiniCPM4.1: Ultra-Efficient LLMs on End Devices, achieving 3+ generation speedup on reasoning tasks
TensorFlow 2.x version's Tutorials and Examples, including CNN, RNN, GAN, Auto-Encoders, FasterRCNN, GPT, BERT examples, etc. TF 2.0版入门实例代码,实战教程。
Democratizing Reinforcement Learning for LLMs
Go to https://github.com/pytorch/tutorials - this repo is deprecated and no longer maintained
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Example TensorFlow codes and Caicloud TensorFlow as a Service dev environment.
Massively parallel rigidbody physics simulation on accelerator hardware.
NeRF (Neural Radiance Fields) and NeRF in the Wild using pytorch-lightning
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts…
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
[ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
Doing simple retrieval from LLM models at various context lengths to measure accuracy
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
A comprehensive guide to building RAG-based LLM applications for production.
PyTorch implementation of Soft Actor-Critic (SAC), Twin Delayed DDPG (TD3), Actor-Critic (AC/A2C), Proximal Policy Optimization (PPO), QT-Opt, PointNet..
TruthfulQA: Measuring How Models Imitate Human Falsehoods
JAX (Flax) implementation of algorithms for Deep Reinforcement Learning with continuous action spaces.
Generative Representational Instruction Tuning
Crosslingual Generalization through Multitask Finetuning