-
Max Planck Institute for Software Systems: MPI SWS
- Delhi
-
12:44
(UTC +05:30) - https://aflah02.github.io/
- @Aflah02101
Lists (23)
Sort Name ascending (A-Z)
Evals
Fun Project Ideas
Hate Speech
Interview Prep
Kernels
Lexicons
LLM API Powertools
LLM Deployment
LLM-RL
LM Eval
Local LLMs
ML PowerTools
Multimodal
Pretraining Dataset
Python Tooling
RAG
RIdeas
Streamlit Components
Style Transfer
BaselinesSynthetic Data
Telemetry
TTS
Websites
Stars
- All languages
- Assembly
- Astro
- Batchfile
- BibTeX Style
- C
- C#
- C++
- CSS
- Cuda
- Cython
- D
- Dart
- Dockerfile
- Fortran
- Go
- Groovy
- HTML
- Java
- JavaScript
- Jsonnet
- Jupyter Notebook
- Kotlin
- LLVM
- Lean
- Lua
- Luau
- MDX
- MLIR
- Makefile
- Markdown
- Mathematica
- Mojo
- NSIS
- Nim
- Nix
- Odin
- PLpgSQL
- Perl
- Python
- R
- Rocq Prover
- Roff
- Ruby
- Rust
- SCSS
- SWIG
- Scala
- Shell
- Svelte
- Swift
- SystemVerilog
- TeX
- TypeScript
- Typst
- Vue
- Zig
- reStructuredText
Explorations into the proposed SDFT, Self-Distillation Enables Continual Learning, from Shenfeld et al. of MIT
A template for research projects in computer science/machine learning using python and julia
Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support
Training library for Megatron-based models with bidirectional Hugging Face conversion capability
For releasing code related to compression methods for transformers, accompanying our publications
[ICML 2024] Official Implementation of SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks
Compressed LLMs for Efficient Text Generation [ICLR'24 Workshop]
A CLI to estimate inference memory requirements for Hugging Face models, written in Python.
Agent skills for Manim to create 3Blue1Brown style animations.
A Model Agnostic function to directly remove specified layers from the LLM
Official implementation of the ICLR paper "Streamlining Redundant Layers to Compress Large Language Models"
Official repository for EMNLP2025 paper "IG-Pruning: Input-Guided Block Pruning for Large Language Models"
Setup guide for ML training on NVIDIA DGX Spark (GB10 Blackwell, CUDA 13, aarch64)
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
Extending the Context of Pretrained LLMs by Dropping Their Positional Embedding
nanoRLHF: from-scratch journey into how LLMs and RLHF really work.
🦋 An Infographic Generation and Rendering Framework, bring words to life with AI!
A compact implementation of SGLang, designed to demystify the complexities of modern LLM serving systems.
Tooling for exact and MinHash deduplication of large-scale text datasets