-
National University of Singapore
- Singapore
- yueliu1999.github.io
Lists (1)
Sort Name ascending (A-Z)
Stars
[NeurIPS 2025] SECA: Semantically Equivalent and Coherent Attacks for Eliciting LLM Hallucinations
MCPMark is a comprehensive, stress-testing MCP benchmark designed to evaluate model and agent capabilities in real-world MCP use.
A benchmark for LLMs on complicated tasks in the terminal
[ICCV 2025] Official PyTorch Implementation of "Curve-Aware Gaussian Splatting for 3D Parametric Curve Reconstruction""
[ICCV 2025] Official PyTorch Implementation of "Learning Self-supervised Part-aware 3D Hybrid Representations of 2D Gaussians and Superquadrics"
[EMNLP2025] From Automation to Autonomy: A Survey on Large Language Models in Scientific Discovery
[arXiv 25] Aesthetics is Cheap, Show me the Text: An Empirical Evaluation of State-of-the-Art Generative Models for OCR
open-source coding LLM for software engineering tasks
SynthRL: Scaling Visual Reasoning with Verifiable Data Synthesis
MLR-Bench: Evaluating AI Agents on Open-Ended Machine Learning Research
🤖️ A collection of papers, blogs and projects of research agents.
AudioTrust: Benchmarking the Multi-faceted Trustworthiness of Audio Large Language Models
A collection of resources and papers on AI Scientist / Robot Scientist
Optimizing Anytime Reasoning via Budget Relative Policy Optimization
The official implementation of the work "Can Indirect Prompt Injection Attacks Be Detected and Removed?"
The official implementation of the work "Defense Against Prompt Injection Attack by Leveraging Attack Techniques"
[NeurIPS 2025] An official source code for paper "GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning".
Official code of paper "Beyond 'Aha!': Toward Systematic Meta-Abilities Alignment in Large Reasoning Models"
Official implementation of MAS-GPT: Training LLMs to Build LLM-based Multi-Agent Systems
Official repository for "Safety in Large Reasoning Models: A Survey" - Exploring safety risks, attacks, and defenses for Large Reasoning Models to enhance their security and reliability.
Awesome-Efficient-Inference-for-LRMs is a collection of state-of-the-art, novel, exciting, token-efficient methods for Large Reasoning Models (LRMs). It contains papers, codes, datasets, evaluation…