-
Tsinghua University - CS
- Beijing
- https://jity16.github.io/
- @jtinyng1
Stars
Claude Code is an agentic coding tool that lives in your terminal, understands your codebase, and helps you code faster by executing routine tasks, explaining complex code, and handling git workflo…
Visual Skills Pack for Obsidian: generate Canvas, Excalidraw, and Mermaid diagrams from text with Claude Code
Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞
JAX (Flax) implementation of algorithms for Deep Reinforcement Learning with continuous action spaces.
An open source quadruped robot pet framework for developing Boston Dynamics-style four-legged robots that are perfect for STEM, coding & robotics education, IoT robotics applications, AI-enhanced r…
[ICRA'25] H2O+: An Improved Framework for Hybrid Offline-and-Online RL with Dynamics Gaps
A simple tool to update bib entries with their official information (e.g., DBLP or the ACL anthology).
Official PyTorch implementation of BAC in the paper "Seizing Serendipity: Exploiting the Value of Past Success in Off-Policy Actor-Critic"
Official PyTorch implementation of "ACE:Off-Policy Actor-Critic with Causality-Aware Entropy Regularization"
Policy Expansion for Bridging Offline-to-Online Reinforcement Learning (ICLR23)
A PyTorch implementation of Implicit Q-Learning
A Python interface for reinforcement learning environments
Standalone library of frequently-used wrappers for dm_env environments.
RL code for training piano-playing policies for RoboPianist.
A Python 3 package for learning Bayesian Networks (DAGs) from data. Official implementation of the paper "DAGMA: Learning DAGs via M-matrices and a Log-Determinant Acyclicity Characterization"
🎓 Hugo Academic Theme 创建一个学术网站. Easily create a beautiful academic résumé or educational website using Hugo, GitHub, and Netlify.
FinRL®: Financial Reinforcement Learning. 🔥
SCOPE-RL: A python library for offline reinforcement learning, off-policy evaluation, and selection
Code accompanying the paper Adversarially Trained Actor Critic for Offline Reinforcement Learning by Ching-An Cheng*, Tengyang Xie*, Nan Jiang, and Alekh Agarwal.
Author's PyTorch implementation of TD7 for online and offline RL
Official Pytorch Implementation of CMLO in the paper ”When to Update Your Model: Constrained Model-based Reinforcement Learning“
a.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.
A curated list of awesome exploration RL resources (continually updated)
Use ChatGPT to summarize the arXiv papers. 全流程加速科研,利用chatgpt进行论文全文总结+专业翻译+润色+审稿+审稿回复
A customizable framework to create maze and gridworld environments