-
Brookhaven National Laboratory
- https://xihaier.github.io/
Highlights
- Pro
Stars
Transforms complex documents like PDFs into LLM-ready markdown/JSON for your Agentic workflows.
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous …
Open-Sora: Democratizing Efficient Video Production for All
A generative world for general-purpose robotics & embodied AI learning.
Fully open reproduction of DeepSeek-R1
An LLM agent that conducts deep research (local and web) on any given topic and generates a long report with citations.
Download your Spotify playlists and songs along with album art and metadata (from YouTube if a match is found).
Intelligent automation and multi-agent orchestration for Claude Code
Lets make video diffusion practical!
This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.
HunyuanVideo: A Systematic Framework For Large Video Generation Model
Enjoy the magic of Diffusion models!
A Library for Advanced Deep Time Series Models for General Time Series Analysis.
Open-source implementation of AlphaEvolve
🚀 Efficient implementations of state-of-the-art linear attention models
A PyTorch library for implementing flow matching algorithms, featuring continuous and discrete flow matching implementations. It includes practical examples for both text and image modalities.
A curated list of resources for using LLMs to develop more competitive grant applications.
MambaOut: Do We Really Need Mamba for Vision? (CVPR 2025)
Official PyTorch implementation of One-Minute Video Generation with Test-Time Training
Open-source deep-learning framework for building, training, and fine-tuning deep learning models using state-of-the-art Physics-ML methods
[ECCV 2024 Oral] LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation.
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
[ICLR & NeurIPS 2025] Repository for Show-o series, One Single Transformer to Unify Multimodal Understanding and Generation.
Official repository for our work on micro-budget training of large-scale diffusion models.
[ICLR'25 Oral] Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think
[NeurIPS 2025] TTRL: Test-Time Reinforcement Learning