-
University of Tübingen
- https://www.katrinrenz.de/
- @KatrinRenz
Highlights
- Pro
Stars
The repository provides code for running inference and finetuning with the Meta Segment Anything Model 3 (SAM 3), links for downloading the trained model checkpoints, and example notebooks that sho…
PlanT 2.0: Exposing Biases and Structural Flaws in Closed-Loop Driving
MDPO: Overcoming the Training-Inference Divide of Masked Diffusion Language Models
[CoRL 2025] CaRL: Learning Scalable Planning Policies with Simple Rewards
[CVPR 2025, Spotlight] SimLingo (CarLLava): Vision-Only Closed-Loop Autonomous Driving with Language-Action Alignment
[CVPR 2025] Volumetric Surfaces: Representing Fuzzy Geometries with Layered Meshes
Official repository for paper "Can LVLMs Obtain a Driver’s License? A Benchmark Towards Reliable AGI for Autonomous Driving"
world modeling challenge for humanoid robots
[WACV 2024 Survey Paper] Multimodal Large Language Models for Autonomous Driving
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
A curated list of awesome LLM/VLM/VLA/World Model for Autonomous Driving(LLM4AD) resources (continually updated)
PyTorch implementation for the paper "Driving with LLMs: Fusing Object-Level Vector Modality for Explainable Autonomous Driving"
[ECCV 2024 Oral] DriveLM: Driving with Graph Visual Question Answering
4DHumans: Reconstructing and Tracking Humans with Transformers
The official PyTorch implementation of the paper "MotionGPT: Finetuned LLMs are General-Purpose Motion Generators"
[ICCV'23] Hidden Biases of End-to-End Driving Models & A starter kit for the CARLA leaderboard 2.0.
Official Code for DragGAN (SIGGRAPH 2023)
A playbook for systematically maximizing the performance of deep learning models.
[CoRL'22] PlanT: Explainable Planning Transformers via Object-Level Representations
Implementation of "Audio Retrieval with Natural Language Queries: A Benchmark Study".
This repository contains the code for our CVPR 2022 paper on "Audio-visual Generalised Zero-shot Learning with Cross-modal Attention and Language"