-
WXG
- Guangzhou
- https://zhuanlan.zhihu.com/data-competitions
Stars
A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171
学习ChatGLM3模型和LangChain框架的架构与核心功能,并基于LangChain+ChatGLM3实现本地知识库问答。
Large World Model -- Modeling Text and Video with Millions Context
Virtual whiteboard for sketching hand-drawn like diagrams
🐙 Guides, papers, lessons, notebooks and resources for prompt engineering, context engineering, RAG, and AI Agents.
HuatuoGPT, Towards Taming Language Models To Be a Doctor. (An Open Medical GPT)
WebGLM: An Efficient Web-enhanced Question Answering System (KDD 2023)
《构筑大语言模型应用:应用开发与架构设计》一本关于 LLM 在真实世界应用的开源电子书,介绍了大语言模型的基础知识和应用,以及如何构建自己的模型。其中包括Prompt的编写、开发和管理,探索最好的大语言模型能带来什么,以及LLM应用开发的模式和架构设计。
jalammar / jalammar.github.io
Forked from barryclark/jekyll-nowBuild a Jekyll blog in minutes, without touching the command line.
Repo for BenTsao [original name: HuaTuo (华驼)], Instruction-tuning Large Language Models with Chinese Medical Knowledge. 本草(原名:华驼)模型仓库,基于中文医学知识的大语言模型指令微调
Repo for Chinese Medical ChatGLM 基于中文医学知识的ChatGLM指令微调
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
2022搜狐校园算法大赛NLP赛道第一名开源方案(实验代码)
[NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
2019CCF-BDCI大赛 最佳创新探索奖获得者 基于OCR身份证要素提取赛题冠军 天晨破晓团队 赛题源码
Source code for "Parsing All: Syntax and Semantics, Dependencies and Spans" published at Findings of EMNLP 2020
WSDM2022留存预测挑战赛 第1名解决方案