Skip to content
Change the repository type filter

Forks

    Repositories list

    • Quantized Attention achieves speedup of 2-5x and 3-11x compared to FlashAttention and xformers, without lossing end-to-end metrics across language, image, and v…
      Cuda
      Apache License 2.0
      389200Updated Jul 8, 2025Jul 8, 2025
    • Implementation of Flash Attention in Jax
      Python
      MIT License
      25900Updated Jul 17, 2024Jul 17, 2024
    • nanobind

      Public
      nanobind: tiny and efficient C++/Python bindings
      C++
      BSD 3-Clause "New" or "Revised" License
      295200Updated Mar 15, 2024Mar 15, 2024
    • 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
      Python
      Apache License 2.0
      33k600Updated Feb 14, 2024Feb 14, 2024
    • 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
      Python
      Apache License 2.0
      33k5600Updated Sep 1, 2023Sep 1, 2023
    • jaxtorch

      Public
      Python
      MIT License
      9100Updated Jun 7, 2023Jun 7, 2023
    • einops

      Public
      Deep learning operations reinvented (for pytorch, tensorflow, jax and others)
      Python
      MIT License
      396700Updated Dec 29, 2022Dec 29, 2022
    • Fast and memory-efficient exact attention
      Python
      BSD 3-Clause "New" or "Revised" License
      2.6k400Updated Dec 17, 2022Dec 17, 2022
    ProTip! When viewing an organization's repositories, you can use the props. filter to filter by custom property.