Efficient, Flexible, Multi-task Batch SFT Data Labeling Tool
-
Updated
Aug 18, 2025 - Python
Efficient, Flexible, Multi-task Batch SFT Data Labeling Tool
Pretrain, finetune ANY AI model of ANY size on multiple GPUs, TPUs with zero code changes.
100000-Instruction-Following-Evaluation-SFT-for-Chinese-LLM-Text-Data
A React-based tool for constructing fine-tuning datasets with list and grid forms, featuring the ability to download and upload data as JSONL files. This project leverages the react-declarative library to create dynamic, interactive forms for defining user inputs, preferred outputs, and non-preferred outputs, along with associated tools
Supervised Finetuning on SmolLM model using the smalltalk dataset.
Gemma3 4B LLM Fine Tuned on 100K Doctor-Patient QA Dataset
A modern Python-based MySQL backup tool with flexible archiving, multi-channel notifications (Telegram, Email, SMS, etc.), remote uploads (SFTP, FTP, SCP), and robust configuration validation.
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
Adapting LLM to the medical domain through SFT, RAG, and multistep fine-tuning to enhance domain knowledge and performance.
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
A powerful tool for creating fine-tuning datasets for LLM
Repo to serve as a baseline/guide for performing post training(SFT/RLHF) of modern LLM models, and evaluating them with baseline datasets.
Wrapped Megatron: As User-Friendly as HuggingFace, As Powerful as Megatron-LM | Megatron封装:和HuggingFace一样方便,和Megatron-LM一样强大
Training Qwen3 to solve Wordle using SFT and GRPO
Add a description, image, and links to the sft topic page so that developers can more easily learn about it.
To associate your repository with the sft topic, visit your repo's landing page and select "manage topics."