Skip to content
View ThomAub's full-sized avatar
  • Paris

Block or report ThomAub

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Starred repositories

11 stars written in Cuda
Clear filter

LLM training in simple, raw C/CUDA

Cuda 29,537 3,515 Updated Jun 26, 2025

A massively parallel, optimal functional runtime in Rust

Cuda 11,225 436 Updated Nov 21, 2024

Tile primitives for speedy kernels

Cuda 3,312 275 Updated Apr 8, 2026

Mirage Persistent Kernel: Compiling LLMs into a MegaKernel

Cuda 2,187 193 Updated Apr 11, 2026

[MICRO'23, MLSys'22] TorchSparse: Efficient Training and Inference Framework for Sparse Convolution on GPUs.

Cuda 1,457 187 Updated Feb 24, 2025

Efficient GPU kernels for block-sparse matrix multiplication and convolution

Cuda 1,065 198 Updated Jun 8, 2023

cuVS - a library for vector search and clustering on the GPU

Cuda 732 180 Updated Apr 10, 2026

High-speed GEMV kernels, at most 2.7x speedup compared to pytorch baseline.

Cuda 128 8 Updated Jul 13, 2024

PyTorch half precision gemm lib w/ fused optional bias + optional relu/gelu

Cuda 78 4 Updated Dec 3, 2024

FlashInfer: Kernel Library for LLM Serving

Cuda 4 Updated Apr 17, 2024