-
Moonshot AI Singapore
- A seat that sees all tourists taking photos with Merlion
-
00:22
(UTC +08:00) - teowu.github.io
- @HaoningTimothy
Lists (1)
Sort Name ascending (A-Z)
Stars
Arena-Hard-Auto: An automatic LLM benchmark.
The official repository of the dots.vlm1 instruct models proposed by rednote-hilab.
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Kimi K2 is the large language model series developed by Moonshot AI team
An open-source AI agent that brings the power of Gemini directly into your terminal.
open-source coding LLM for software engineering tasks
A benchmark for evaluating vision-centric, complex video reasoning.
Kimi-VL: Mixture-of-Experts Vision-Language Model for Multimodal Reasoning, Long-Context Understanding, and Strong Agent Capabilities
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
Turn any PDF or image document into structured data for your AI. A powerful, lightweight OCR toolkit that bridges the gap between images/PDFs and LLMs. Supports 100+ languages.
PyTorch code for our paper "Grounding-IQA: Grounding Multimodal Language Model for Image Quality Assessment"
[ACMMM2025] Official released code for VQA² series models
[CVPR 2025] Official Dataloader and Evaluation Scripts for VideoAutoArena.
[CVPR 2025] Official Dataloader and Evaluation Scripts for VideoAutoBench.
MathVista: data, code, and evaluation for Mathematical Reasoning in Visual Contexts
[ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, Lei Li, Sishuo Chen, Xu Sun, Lu Hou
[Nips 2025] EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation
A Versatile Video-LLM for Long and Short Video Understanding with Superior Temporal Localization Ability
VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs
🔥🔥MLVU: Multi-task Long Video Understanding Benchmark
PyTorch code for our paper "Dog-IQA: Standard-guided Zero-shot MLLM for Mix-grain Image Quality Assessment"
Codebase for Aria - an Open Multimodal Native MoE
Code for DeCo: Decoupling token compression from semanchc abstraction in multimodal large language models
[ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning