Skip to content

tongnie/awesome-llm4tr

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

55 Commits
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Exploring the Roles of Large Language Models in Reshaping Transportation Systems: A Survey, Framework, and Roadmap

DOI Badge arXiv Badge Stars Badge

The LLM4TR framework proposed in this survey.

A curated collection of papers, datasets, and resources related to Large Language Models for Transportation (LLM4TR).

This repository serves as the online companion to our survey "Exploring the Roles of Large Language Models in Reshaping Transportation Systems: A Survey, Framework, and Roadmap", Artificial Intelligence for Transportation (AIT), 1, 100003. Paper page

๐Ÿ“ข The rapid advancement of Large Language Models (LLMs) is creating a paradigm shift in transportation research, moving from purely data-driven models to knowledge-infused intelligent systems. This repository aims to track the cutting-edge developments in this exciting field, serving as a hub for researchers, practitioners, and students. Stay tuned!

Feel free to contact us if you have any suggestions or would like to discuss with us by e-mail: tong.nie@connect.polyu.hk, wei.w.ma@polyu.edu.hk

Table of Contents

๐Ÿ”ฅ News & Updates

  • ๐ŸŽ‰ [June 2025] Official Publication! Our survey has been published in the inaugural issue of Artificial Intelligence for Transportation (AIT), a new flagship journal led by the Chinese Overseas Transportation Association (COTA). As the third paper in its first volume (1, 100003), we are honored to contribute to this exciting new venue for AI in transportation research!
  • โœ… [March 2025] ArXiv Version: Our survey is released on arXiv!
  • โœ… [March 2025] Repository Launch: The awesome-llm4tr repository is initialized.

๐Ÿ“ Cite Our Work

If you find our survey or this repository useful for your research, please cite our paper:

@article{nie2025exploring,
  title={Exploring the roles of large language models in reshaping transportation systems: A survey, framework, and roadmap},
  author={Nie, Tong and Sun, Jian and Ma, Wei},
  journal={Artificial Intelligence for Transportation},
  volume={1},
  pages={100003},
  year={2025},
  publisher={Elsevier},
  doi={10.1016/j.ait.2025.100003}
}

๐Ÿ“– Framework and Taxonomy

Definition: LLM4TR refers to the methodological paradigm that systematically harnesses emergent capabilities of LLMs to enhance transportation tasks through four synergistic roles: transforming raw data into understandable insights, distilling domain-specific knowledge into computable structures, synthesizing adaptive system components, and orchestrating optimal decisions. We survey the existing literature and summarize how LLMs are exploited to solve transportation problems from a methodological perspective, i.e., the roles of LLMs in transportation systems. They generally include four aspects:

The literature classification procedure in this survey.

๐Ÿ“š Papers by Category

Here, we list representative papers according to the four roles defined in the LLM4TR framework.

LLMs as Information Processors

Function: Process and fuse heterogeneous transportation data from multiple sources (text, sensor data, user feedback) through contextual encoding, analytical reasoning, and multimodal integration.

LLMs as Information Processors.

LLMs as Knowledge Encoders

Function: Extract and formalize transportation domain knowledge from unstructured data through explicit rule extraction and latent semantic embedding.

LLMs as Knowledge Encoders.

LLMs as Component Generators

Function: Create functional algorithms, synthetic environments, and evaluation frameworks through instruction-followed content generation.

LLMs as Component Generators.

LLMs as Decision Facilitators

Function: Predict traffic dynamics, optimize decisions, and simulate human-like reasoning, establishing new paradigms as generalized task solvers.

LLMs as Decision Facilitators.

๐Ÿ“ˆ Research Trend

An overview of the current research trends, visualized according to our taxonomy.

Heatmap of the current research trend and pie chart of the proportion of the four roles of LLMs in different tasks.

โญ Overview of Mainstream LLMs

Model Release Date Organization Size (B) Data (TB) Hardware Cost Public Access
T5 2019.10 Google 11 750 GB of text 1024 TPU v3 Yes
GPT-3 2020.5 OpenAI 175 300 B tokens - No
PaLM 2022.4 Google 540 780 B tokens 6144 TPU v4 No
LLaMA 2023.2 Meta 65 1.4 T tokens 2048 A100 GPU Partial
GPT-4 2023.3 OpenAI - - - No
LLaMA-2 2023.7 Meta 70 2 T tokens 2000 A100 GPU Yes
Mistral-7B 2023.9 Mistral AI 7 - - Yes
Qwen-72B 2023.11 Alibaba 72 3 T tokens - Yes
Grok-1 2024.3 xAI 314 - - Yes
Claude 3 2024.3 Anthropic - - - No
ChatGLM-4 2024.6 Zhipu AI 9 10 T tokens - Yes
LLaMA-3.1 2024.7 Meta 405 15 T tokens 16k H100 GPU Yes
Gemma-2 2024.6 Google 27 13 T tokens 6144 TPUv5p Yes
DeepSeek-V2 2024.12 DeepSeek 671 (MoE) 14.8 T tokens 2048 H800 GPU Yes

๐Ÿ’ก Foundational LLM Techniques

This section lists core papers on the foundational techniques of LLMs, providing essential context for their application in transportation. It is designed to be a comprehensive resource for understanding how LLMs work.

1. Core Architectures & Concepts

2. Seminal Pre-trained Models

3. Instruction Tuning & Alignment

4. Parameter-Efficient Fine-Tuning (PEFT)

5. Advanced Reasoning & Prompting

6. Integrating External Knowledge & Tools

7. Efficiency & Optimization

8. Key Evaluation Benchmarks


๐Ÿ“‹ Summary of Language-Enhanced Datasets

Dataset Year Venue Task Use Case in LLM Development
KITTI 2012 CVPR 3D object detection, tracking, stereo, and visual odometry Foundational benchmark for classic perception tasks in autonomous driving.
Waymo Open Dataset 2019 CVPR Perception (detection, tracking) and motion prediction Large-scale, high-resolution sensor data for training and evaluating AD models.
BDD-X 2018 ECCV Action interpretation and control signal prediction Explainable end-to-end driving through visual question answering.
SUTD-TrafficQA 2021 CVPR Video causal reasoning over traffic events Evaluating the reasoning capability over 6 tasks.
Argoverse 2 2023 CVPR Motion forecasting, lidar object detection Contains sensor data and HD maps with a focus on prediction and forecasting tasks.
TrafficSafety-2K 2023 arXiv Annotated traffic incident and crash report analysis GPT fine-tuning for safety situational awareness.
NuPrompt 2023 AAAI Object-centric language prompt set for 3D driving scenes Prompt-based driving task to predict the described object trajectory.
Talk2Car 2019 ICCV Referring object localization via natural language Grounding language commands to objects in a visual scene for human-vehicle interaction.
LaMPilot 2024 CVPR Code generation for autonomous driving decisions CoT reasoning and instruction following for lane changes and speed adaptation.
CoVLA 2024 arXiv Vision-Language-Action alignment (80+ hrs driving videos) Trajectory planning with natural language maneuver descriptions.
VLAAD 2024 WACV Natural language description of driving scenarios QA systems for driving situation understanding.
CrashLLM 2024 arXiv Crash outcome prediction (severity, injuries) What-if causal analysis for traffic safety using 19k crash reports.
TransportBench 2024 arXiv Answering undergraduate-level transportation engineering problem Benchmarking LLMs on planning, design, management, and control questions.
Driving QA 2024 ICRA 160k driving QA pairs with control commands Interpreting scenarios, answering questions, and decision-making.
MAPLM 2024 CVPR Multimodal traffic scene dataset including context, image, point cloud, and HD map Visual instruction-tuning LLMs and VLMs and vision QA tasks.
DrivingDojo 2024 NeurIPS Video clips with maneuvers, multi-agent interplay, and driving knowledge Training and action instruction following benchmark for driving world models.
TransportationGames 2024 arXiv Benchmarks of LLMs in memorizing, understanding, and applying transportation knowledge Grounding (M)LLMs in transportation-related tasks.
NuScenes-QA 2024 AAAI Benchmark for vision QA in autonomous driving, with 34K visual scenes and 460K QA pairs Developing 3D detection and VQA techniques for end-to-end autonomous driving systems.
TUMTraffic-VideoQA 2025 arXiv Temporal traffic video understanding Benchmarking video reasoning for multiple-choice video question answering.
V2V-QA 2025 arXiv Cooperative perception via V2V communication Fuse perception information from multiple CAVs and answer driving-related questions.
DriveBench 2025 arXiv A comprehensive benchmark of VLMs for perception, prediction, planning, and explanation Visual grounding and multi-modal understanding for autonomous driving.

๐Ÿ“„ Representative Surveys on LLMs

Paper Title Year Venue Scope and Focus
A survey of Large Language Models 2023 arXiv Reviews the evolution of LLMs, pretraining, adaptation, post-training, evaluation, and benchmarks.
Large Language Models: A Survey 2024 arXiv Reviews LLM families (GPT, LLaMA, PaLM), training techniques, datasets, and benchmark performance.
Retrieval-Augmented Generation for Large Language Models: A Survey 2023 arXiv Introduces the progress of RAG paradigms, including the naive RAG, the advanced RAG, and the modular RAG.
A Survey on In-context Learning 2022 arXiv Summarizes training strategies, prompt designing strategies, and various ICL application scenarios.
Instruction Tuning for Large Language Models: A Survey 2023 arXiv Reviews methodology of SFT, SFT datasets, applications to different modalities, and influence factors.
Towards Reasoning in Large Language Models: A Survey 2022 ACL Examines techniques for improving and eliciting reasoning in LLMs, methods and benchmarks for evaluating reasoning abilities.
A Survey of LLM Surveys 2024 GitHub Compiles 150+ surveys across subfields like alignment, safety, and multimodal LLMs.

๐Ÿ› ๏ธ Popular Open-Source Libraries for LLM Development

Library Name Basic Functions Use Cases
Hugging Face Transformers Pretrained models (NLP, vision) and fine-tuning pipelines Model deployment, adapt tuning
DeepEval Framework for evaluating LLM outputs using metrics like groundedness and bias Educational applications, hallucination detection
RAGAS Quantifies RAG pipeline performance Context relevance scoring, answer quality
Sentence Transformers Generates dense text embeddings for semantic similarity tasks Survey item correlation analysis, retrieval
LangChain Chains LLM calls with external tools for multi-step workflows RAG, agentic reasoning, data preprocessing
DeepSpeed A deep learning optimization library from Microsoft for training large-scale models Distributed training, memory optimization, pipeline parallelism
FastMoE A specialized training library for Mixture-of-Experts (MoE) models based on PyTorch Transfer Transformer models to MoE models, data parallelism, model parallelism
Ollama Serve and run large language models locally Offline inference, privacy-sensitive apps, local development
OpenLLM An open platform for operating LLMs in production Scalable model serving, cloud/on-prem hosting

๐Ÿ–ฅ๏ธ Hardware Requirements for Fine-Tuning

A rough estimate of hardware requirements for fine-tuning and inference across LLaMA model sizes.

BS = Batch Size. Estimated values marked (est.) derive from scaling laws. Inference rates are measured at batch size 1 unless noted. Actual requirements may differ.

Model Size Full Tuning GPUs LoRA Tuning GPUs Full Tuning BS/GPU LoRA BS/GPU Tuning Time (Hours) Inference Rate (Tokens/s)
7B 2ร—A100 80GB 1ร—RTX 4090 24GB 1-2 4-8 3-5 27-30
13B 4ร—A100 80GB (est.) 2ร—A100 40GB 1 2-4 8-12 18-22
70B 8ร—H200 80GB 4ร—H200 80GB 1 1-2 24-36 12-15
405B 64ร—H200 80GB (est.) 16ร—H200 80GB (est.) 1 (est.) 1 (est.) 72-96 (est.) 5-8

๐ŸŒŸ Awesome Lists and Resource Hubs

A curated list of other high-quality GitHub repositories and resource collections relevant to LLMs, autonomous driving, and AI in transportation.

Repository Area Description
ge25nab/Awesome-VLM-AD-ITS VLM, AD, ITS A focused collection of papers on Vision-Language Models for Autonomous Driving and ITS.
kaushikb11/awesome-llm-agents LLM Agents A comprehensive list of papers, frameworks, and resources for building and understanding LLM-based agents.
RUCAIBox/LLMSurvey LLM Survey An extensive survey of Large Language Models, covering papers on pre-training, fine-tuning, and reasoning.
BradyFU/Awesome-Multimodal-Large-Language-Models MLLMs A curated list of resources for Multimodal Large Language Models (MLLMs), including papers and code.
manfreddiaz/awesome-autonomous-driving Autonomous Driving A massive list of resources for all aspects of autonomous driving, from sensors to planning.
NiuTrans/ABigSurveyOfLLMs LLM Survey A meta-survey compiling over 150 surveys on LLMs across various subfields like alignment and safety.
thunlp/GNNPapers Graph Neural Networks A must-read list of papers on Graph Neural Networks, highly relevant for transportation network modeling.
huggingface/datasets Datasets The official repository for thousands of datasets, easily accessible for training and evaluation.
microsoft/DeepSpeed LLM Training A deep learning optimization library that makes large model training and inference more efficient.
bentoml/OpenLLM LLM Deployment An open platform for operating LLMs in production, simplifying deployment and scaling.
jbhuang0604/awesome-computer-vision Computer Vision A classic, curated list of essential resources for all things computer vision.
diff-usion/Awesome-Diffusion-Models Diffusion Models A collection of resources for diffusion models, crucial for generative tasks like world simulation.

๐Ÿค How to Contribute

We warmly welcome contributions from the community! If you have found a relevant paper, code, or resource that we missed, please feel free to:

  • Open an Issue to suggest an addition or report an error.
  • Create a Pull Request to directly add your content. Please follow the existing format to ensure consistency.

Let's build the best resource for LLM4TR together!

License

This repository is released under the MIT LICENSE.

About

Exploring the Roles of Large Language Models in Reshaping Transportation Systems: A Survey, Framework, and Roadmap

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published