-
ShanghaiTech University
- Shanghai
-
03:41
(UTC +08:00)
Highlights
- Pro
Starred repositories
📚 Freely available programming books
微舆:人人可用的多Agent舆情分析助手,打破信息茧房,还原舆情原貌,预测未来走向,辅助决策!从0实现,不依赖任何框架。
Opiniated RAG for integrating GenAI in your apps 🧠 Focus on your product rather than the RAG. Easy integration in existing products with customisation! Any LLM: GPT4, Groq, Llama. Any Vectorstore: …
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and…
Official Code for DragGAN (SIGGRAPH 2023)
Open-Sora: Democratizing Efficient Video Production for All
A generative world for general-purpose robotics & embodied AI learning.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"
A lightweight, powerful framework for multi-agent workflows
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
The pytest framework makes it easy to write small tests, yet scales to support complex functional testing
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
🐍 Geometric Computer Vision Library for Spatial AI
Implementation of Nougat Neural Optical Understanding for Academic Documents
ImageBind One Embedding Space to Bind Them All
Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion.
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
AlphaFold 3 inference pipeline.
Utilities intended for use with Llama models.
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
A concise but complete full-attention transformer with a set of promising experimental features from various papers