-
Shanghai Jiao Tong University
- Shanghai
-
03:19
(UTC -12:00) - https://www.sjtu.edu.cn/
Highlights
- Pro
Lists (1)
Sort Name ascending (A-Z)
Stars
Belief in the Machine: Investigating Epistemological Blind Spots of Language Models
Code for the NPJ AI paper "How Large Language Models Encode Theory-of-Mind: A Study on Sparse Parameter Patterns"
ToMCAT: Theory of Mind-based Cognitive Architecture for Teams
tomsup 👍 Theory of Mind Simulation using Python. A package that allows for easy agent-based modelling of recursive Theory of Mind
AI tutor powered by Theory-of-Mind reasoning
Machine Theory of Mind Reading List. Built upon EMNLP Findings 2023 Paper: Towards A Holistic Landscape of Situated Theory of Mind in Large Language Models
An open platform for enhancing the capability of LLMs in workflow orchestration.
Odyssey: Empowering Minecraft Agents with Open-World Skills
LangGPT: Empowering everyone to become a prompt expert! 🚀 📌 结构化提示词(Structured Prompt)提出者 📌 元提示词(Meta-Prompt)发起者 📌 最流行的提示词落地范式 | Language of GPT The pioneering framework for structured & meta-prompt…
LangChain & Prompt Engineering tutorials on Large Language Models (LLMs) such as ChatGPT with custom data. Jupyter notebooks on loading and indexing data, creating prompt templates, CSV agents, and…
Explain Before You Answer: A Survey on Compositional Visual Reasoning
Official codebase for the paper "Training-Free Personalization via Retrieval and Reasoning on Fingerprints"
Official Implementation for "MyVLM: Personalizing VLMs for User-Specific Queries" (ECCV 2024)
This repo aims to record resource of role-playing abilities in LLMs, including dataset, paper, application, etc.
《大模型白盒子构建指南》:一个全手搓的Tiny-Universe
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
[CVPR 2025] RAP: Retrieval-Augmented Personalization
A curated list of Awesome Personalized Large Multimodal Models resources
CheX-Phi3.5V is a vision-language model (VLM) for chest X-ray interpretation.
This repository contains the official code for our paper: Thinking Before Looking: Improving Multimodal LLM Reasoning via Mitigating Visual Hallucination
Multimodal Chain-of-Thought Reasoning: A Comprehensive Survey
GMAI-VL & GMAI-VL-5.5M: A Large Vision-Language Model and A Comprehensive Multimodal Dataset Towards General Medical AI.
R1-VL: Learning to Reason with Multimodal Large Language Models via Step-wise Group Relative Policy Optimization
[ 🎯 NAACL 2025 ] MedThink: A Rationale-Guided Framework for Explaining Medical Visual Question Answering
[AAAI-2026] Patho-R1: A Multimodal Reinforcement Learning-Based Pathology Expert Reasoner
[NIPS 2025] Chiron-o1: Igniting Multimodal Large Language Models towards Generalizable Medical Reasoning via Mentor-Intern Collaborative Search