Stars
🎉 PILOT: A Pre-trained Model-Based Continual Learning Toolbox
[Paper][AAAI 2023] DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning
[CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want
(TPAMI 2024) A Survey on Open Vocabulary Learning
The official GitHub page for the survey paper "A Survey of Large Language Models".
✨✨Latest Advances on Multimodal Large Language Models
ICCV 2023: CLIPN for Zero-Shot OOD Detection: Teaching CLIP to Say No
Curated tutorials and resources for Large Language Models, AI Painting, and more.
A new framework for open-vocabulary object detection, based on maskrcnn-benchmark
[CVPR 2022] Official code for "RegionCLIP: Region-based Language-Image Pretraining"
[Pattern Recognition 25] CLIP Surgery for Better Explainability with Enhancement in Open-Vocabulary Tasks
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
[ICCV 2023] Code for "Not All Features Matter: Enhancing Few-shot CLIP with Adaptive Prior Refinement"
[CVPR 2023] Unofficial re-implementation of "WinCLIP: Zero-/Few-Shot Anomaly Classification and Segmentation".
This is an official PyTorch code for our accepted paper "When All We Need is a Piece of the Pie: A Generic Framework for Optimizing Two-way Partial AUC" in ICML-2021
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
A PyTorch Library for Meta-learning Research
Cross-modal few-shot adaptation with CLIP
[CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)
Desktop application of new Bing's AI-powered chat (Windows, macOS and Linux)
A library for ML benchmarking. It's powerful.