Skip to content
View leixy76's full-sized avatar

Block or report leixy76

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Starred repositories

12 stars written in Cuda
Clear filter

LLM training in simple, raw C/CUDA

Cuda 28,415 3,332 Updated Jun 26, 2025

DeepEP: an efficient expert-parallel communication library

Cuda 8,812 1,032 Updated Dec 5, 2025

DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling

Cuda 5,971 777 Updated Dec 8, 2025

FlashInfer: Kernel Library for LLM Serving

Cuda 4,289 602 Updated Dec 18, 2025

[ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.

Cuda 2,861 286 Updated Dec 11, 2025

Static suckless single batch CUDA-only qwen3-0.6B mini inference engine

Cuda 534 44 Updated Sep 8, 2025

Reference implementation of Megalodon 7B model

Cuda 527 54 Updated May 17, 2025

flash attention tutorial written in python, triton, cuda, cutlass

Cuda 459 50 Updated May 14, 2025

llama3.cuda is a pure C/CUDA implementation for Llama 3 model.

Cuda 349 26 Updated Apr 27, 2025

A quantization algorithm for LLM

Cuda 146 8 Updated Jun 21, 2024
Cuda 5 Updated Jun 21, 2024