Skip to content
#

fine-tune

Here are 80 public repositories matching this topic...

A LLM-based chatbot can solve math problems, give clearly explainable reasons and self-training by using reinforcement learning, Fine-tune Mistral v0.3 (7B) on MathQA-40K datasets, Use Unsloth framework for increasing performance model, speed training time, Apply Google Mind paper’s techniques for increasing accuracy and performance of model

  • Updated Jul 5, 2025
  • Jupyter Notebook

Implement of Behavior Cloning (BC) and Conservative Q-Learning (CQL) algorithms for training reinforcement learning models using a dataset of state-action pairs. It provides an environment for experimenting with these algorithms, comparing their performance, and analyzing the effects of different parameters and dataset variations on training result

  • Updated Jan 17, 2025
  • Python

This repository contains a PyTorch implementation of a Convolutional Neural Network (CNN) for classifying the MNIST dataset. The project explores different fine-tuning techniques, including LoRA (Low-Rank Adaptation), DoRA (Dynamic Low-Rank Adaptation), and QLoRA (Quantized Low-Rank Adaptation), to improve model performance and efficiency.

  • Updated Mar 10, 2025
  • Python

A Novel Approach for Alzheimer's Classification Utilizing Ensemble Learning on Pre-trained Neural Networks Fine-tuned on Pre-processed and Augmented Alzheimer's Dataset

  • Updated Jul 5, 2025
  • Jupyter Notebook

Improve this page

Add a description, image, and links to the fine-tune topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the fine-tune topic, visit your repo's landing page and select "manage topics."

Learn more