-
MAC Lab, XMU
- Fujian, China
-
06:25
(UTC +08:00)
Highlights
- Pro
Lists (21)
Sort Name ascending (A-Z)
Stars
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
A latent text-to-image diffusion model
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
12 Lessons to Get Started Building AI Agents
Google Research
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
Instruct-tune LLaMA on consumer hardware
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
StableLM: Stability AI Language Models
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
This open-source curriculum introduces the fundamentals of Model Context Protocol (MCP) through real-world, cross-language examples in .NET, Java, TypeScript, JavaScript, Rust and Python. Designed …
强化学习中文教程(蘑菇书🍄),在线阅读地址:https://datawhalechina.github.io/easy-rl/
PyTorch code and models for the DINOv2 self-supervised learning method.
LAVIS - A One-stop Library for Language-Vision Intelligence
Code release for NeRF (Neural Radiance Fields)
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". A…
Reference PyTorch implementation and models for DINOv3
Implementation of Dreambooth (https://arxiv.org/abs/2208.12242) with Stable Diffusion
OpenMMLab Multimodal Advanced, Generative, and Intelligent Creation Toolbox. Unlock the magic 🪄: Generative-AI (AIGC), easy-to-use APIs, awsome model zoo, diffusion models, for text-to-image genera…
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
This is the official code for MobileSAM project that makes SAM lightweight for mobile applications and beyond!