Popular repositories Loading
-
SageAttention
SageAttention PublicForked from thu-ml/SageAttention
Quantized Attention achieves speedup of 2-5x and 3-11x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and video models.
Cuda
-
radial-attention
radial-attention PublicForked from mit-han-lab/radial-attention
Radial Attention Official Implementation
Python
-
Jenga
Jenga PublicForked from dvlab-research/Jenga
Official Implementation: Training-Free Efficient Video Generation via Dynamic Token Carving
Python
-
moon-lite-attention
moon-lite-attention PublicForked from Dao-AILab/flash-attention
Fast and memory-efficient exact attention (fork of flash-attention)
Python
-
Wan2.1-pv-skip
Wan2.1-pv-skip PublicForked from Wan-Video/Wan2.1
Wan: Open and Advanced Large-Scale Video Generative Models
Python
Repositories
- SageAttention Public Forked from thu-ml/SageAttention
Quantized Attention achieves speedup of 2-5x and 3-11x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and video models.
moonmath-ai/SageAttention’s past year of commit activity - moon-lite-attention Public Forked from Dao-AILab/flash-attention
Fast and memory-efficient exact attention (fork of flash-attention)
moonmath-ai/moon-lite-attention’s past year of commit activity - Jenga Public Forked from dvlab-research/Jenga
Official Implementation: Training-Free Efficient Video Generation via Dynamic Token Carving
moonmath-ai/Jenga’s past year of commit activity - radial-attention Public Forked from mit-han-lab/radial-attention
Radial Attention Official Implementation
moonmath-ai/radial-attention’s past year of commit activity - Wan2.1-pv-skip Public Forked from Wan-Video/Wan2.1
Wan: Open and Advanced Large-Scale Video Generative Models
moonmath-ai/Wan2.1-pv-skip’s past year of commit activity
People
This organization has no public members. You must be a member to see who’s a part of this organization.
Top languages
Loading…
Most used topics
Loading…