Stars
TradingAgents: Multi-Agents LLM Financial Trading Framework
Official code of "The Automated but Risky Game: Modeling Agent-to-Agent Negotiations and Transactions in Consumer Markets"
A library for generative social simulation
[ICLR 2025] Official implementation for "SafeWatch: An Efficient Safety-Policy Following Video Guardrail Model with Transparent Explanations"
Repo for the paper "Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks".
AIR-Bench 2024 is a safety benchmark that aligns with emerging government regulations and company policies
Official code for NeurIPS 2025 paper "GRIT: Teaching MLLMs to Think with Images"
[ECCV 2024] API: Attention Prompting on Image for Large Vision-Language Models
[CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding
Unofficial re-implementation of "Trusting Your Evidence: Hallucinate Less with Context-aware Decoding"
[ICLR 2025] This is the official repository of our paper "MedTrinity-25M: A Large-scale Multimodal Dataset with Multigranular Annotations for Medicine“
Repository containing all relevant steps to create visuals and summary statistics for the research project "Prompt Injection Attacks on Large Language Models in Oncology" by Clusmann et al.
Medical o1, Towards medical complex reasoning with LLMs
GMAI-MMBench: A Comprehensive Multimodal Evaluation Benchmark Towards General Medical AI.
This repository implements teleoperation of the Unitree humanoid robot using XR Devices.
This repository compiles a list of papers related to the application of video technology in the field of robotics! Star⭐ the repo and follow me if you like what you see🤩.
[IROS 2025 Award Finalist] The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems
Codes for paper "SafeAgentBench: A Benchmark for Safe Task Planning of \\ Embodied LLM Agents"
This is the official repository for the ICLR 2025 accepted paper Badrobot: Manipulating Embodied LLMs in the Physical World.
Official repo of Exploring the Adversarial Vulnerabilities of Vision-Language-Action Models in Robotics
Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.