Pytorch Lightning implementation for various NLP tasks
This repository contains:
- Finetuning Transformer decoder (Llama/Komodo/Qwen) for Indonesian Regional Language Translations, modes: Vanilla finetuning, Low Rank Adaptation (LoRA) via peft, Fully Sharded Data Parallel (FSDP) with Pytorch XLA.
- Additional pretraining script for Transformer encoder (XLM Roberta) with Masked Language Modelling Objective.
- Finetuning Transformer encoder (XLM Roberta) for zero shot classification.
- Finetuning Transformer encoder (XLM Roberta) for sentiment analysis.