Skip to content
View Wei-Baldwin-Zeng's full-sized avatar

Block or report Wei-Baldwin-Zeng

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Showing results

[CoRL 2025] Repository relating to "TrackVLA: Embodied Visual Tracking in the Wild"

Python 264 18 Updated Oct 16, 2025

RetinaFace: Deep Face Detection Library for Python

Python 1,761 180 Updated Aug 11, 2025
Jupyter Notebook 839 142 Updated Jul 10, 2024

Official Algorithm Implementation of ICML'23 Paper "VIMA: General Robot Manipulation with Multimodal Prompts"

Python 831 96 Updated Apr 18, 2024

Embodied Reasoning Question Answer (ERQA) Benchmark

Python 240 12 Updated Mar 12, 2025

A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites

4,077 318 Updated Oct 17, 2025

This repository provides valuable reference for researchers in the field of multimodality, please start your exploratory travel in RL-based Reasoning MLLMs!

1,251 58 Updated Oct 18, 2025

SAPIEN Manipulation Skill Framework, an open source GPU parallelized robotics simulator and benchmark, led by Hillbot, Inc.

Python 2,216 379 Updated Oct 31, 2025

Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning

Python 75 6 Updated May 17, 2025

MiniCPM-V 4.5: A GPT-4o Level MLLM for Single Image, Multi Image and High-FPS Video Understanding on Your Phone

Python 22,184 1,665 Updated Sep 24, 2025

OpenVLA: An open-source vision-language-action model for robotic manipulation.

Python 4,321 516 Updated Mar 23, 2025

Embodied Chain of Thought: A robotic policy that reason to solve the task.

Python 322 16 Updated Apr 5, 2025

A-MEM: Agentic Memory for LLM Agents

Python 664 79 Updated Oct 21, 2025

Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasks

Python 178 15 Updated Sep 24, 2025

RoboBrain 2.0: Advanced version of RoboBrain. See Better. Think Harder. Do Smarter. πŸŽ‰πŸŽ‰πŸŽ‰

Python 674 57 Updated Sep 30, 2025

[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions

Python 817 48 Updated Aug 21, 2025

This is a curated list of "Embodied AI or robot with Large Language Models" research. Watch this repository for the latest updates! πŸ”₯

1,580 89 Updated Oct 30, 2025

πŸ€– RoboOS: A Universal Embodied Operating System for Cross-Embodied and Multi-Robot Collaboration

Python 236 28 Updated Sep 4, 2025

Official repository of "SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory"

Python 6,986 478 Updated Mar 18, 2025

Full Autonomy Stack for Unitree Go2

C++ 304 43 Updated Apr 1, 2025

State-of-the-art 2D and 3D Face Analysis Project

Python 26,946 5,814 Updated Sep 27, 2025

[RSS 2024 & RSS 2025] VLN-CE evaluation code of NaVid and Uni-NaVid

Python 295 20 Updated Oct 15, 2025
Python 279 34 Updated Mar 17, 2025

[NeurIPS 2024] Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation

Python 6,951 691 Updated Jan 22, 2025

PyTorch implementation of paper "ARTrack" and "ARTrackV2"

Python 292 35 Updated Oct 20, 2025

[AAAI2024]Hybrid-SORT: Weak Cues Matter for Online Multi-Object Tracking

Python 235 30 Updated Apr 2, 2024

BoxMOT: Pluggable SOTA multi-object tracking modules modules for segmentation, object detection and pose estimation models

Python 7,769 1,855 Updated Oct 31, 2025

πŸ” Explore Egocentric Vision: research, data, challenges, real-world apps. Stay updated & contribute to our dynamic repository! Work-in-progress; join us!

120 8 Updated Nov 23, 2024

Official repo and evaluation implementation of VSI-Bench

Python 616 37 Updated Aug 5, 2025
Next