-
Westlake University
- China
- yliu-cs.github.io
Stars
- All languages
- Astro
- Batchfile
- BibTeX Style
- BitBake
- C
- C#
- C++
- CSS
- Cuda
- Dart
- Dockerfile
- Emacs Lisp
- Go
- HTML
- Java
- JavaScript
- Jinja
- Jupyter Notebook
- Kotlin
- Lua
- MATLAB
- MDX
- Makefile
- Markdown
- OpenEdge ABL
- PHP
- Pascal
- Perl
- PowerShell
- Python
- R
- Rich Text Format
- Ruby
- Rust
- SCSS
- Shell
- Smali
- Squirrel
- Swift
- TeX
- Text
- Tree-sitter Query
- TypeScript
- Typst
- Vim Script
- Vue
HiF-VLA: An efficient, bidirectional spatiotemporal expansion Vision-Language-Action Model
A pure Unix shell script ACME client for SSL / TLS certificate automation
[AAAI 2026] Official code for MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Manipulation
[ICLR 2025 Oral] Seer: Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulation
IPTV checker tool for Docker && Desktop && CMD, check your playlist is available
收集全网 Android TV电视盒子应用,涵盖影视、直播、K歌、工具、游戏等类型,整理优质APK资源,支持便捷下载与自动更新。提供安全验证、分类索引与兼容性标注,助力用户打造家庭影音娱乐中心! ✅ TVBox/影视仓等影音壳接口配置源。
📺IPTV直播源自动更新平台『🤖全自动采集、筛选、测速、生成流程🚀』:每日更新两次,提供免费结果订阅,输入播放器即可观看;频道自定义;个性化配置;订阅源/组播源/酒店源/本地源;IPv4/IPv6;接口验效;推流;工作流、Docker、命令行、GUI多端部署 | IPTV live streaming source automatic update platform
油桃TV 电视浏览器 可看各大卫视CCTV直播 无需电视VIP 适配爱奇艺等主流视频平台
Tracking the latest and greatest research papers on diffusion large language models.
Adaptive Multimodal Reasoning via Reinforcement Learning
ContextVLA: Vision-Language-Action Model with Amortized Multi-Frame Context
Does Understanding Inform Generation in Unified Multimodal Models? From Analysis to Path Forward
Official code for 'Learning to Rank for In-Context Example Retrieval'
siiRL: Shanghai Innovation Institute RL Framework for Advanced LLMs and Multi-Agent Systems
Awesome papers for affective computing with llm and mllm
[ICRA'25] Official code repository of "QUART-Online: Latency-Free Large Multimodal Language Model for Quadruped Robot Learning"
The official implementation of Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight
NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-based Preference Rewards
《大模型白盒子构建指南》:一个全手搓的Tiny-Universe
A Survey on Reinforcement Learning of Vision-Language-Action Models for Robotic Manipulation
Official Implementation of "MMaDA-Parallel: Multimodal Large Diffusion Language Models for Thinking-Aware Editing and Generation"
CANDI: Continuous and Discrete Diffusion
RLinf is a flexible and scalable open-source infrastructure designed for post-training foundation models (LLMs, VLMs, VLAs) via reinforcement learning.
Official Implementation for the paper "d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning"
Official implementation of "Continuous Autoregressive Language Models"
🔥 The first open-sourced diffusion vision-langauge-action model.