Skip to content
#

attention

Here are 813 public repositories matching this topic...

Keras beit,caformer,CMT,CoAtNet,convnext,davit,dino,efficientdet,edgenext,efficientformer,efficientnet,eva,fasternet,fastervit,fastvit,flexivit,gcvit,ghostnet,gpvit,hornet,hiera,iformer,inceptionnext,lcnet,levit,maxvit,mobilevit,moganet,nat,nfnets,pvt,swin,tinynet,tinyvit,uniformer,volo,vanillanet,yolor,yolov7,yolov8,yolox,gpt2,llama2, alias kecam

  • Updated Nov 12, 2025
  • Python
annotated_deep_learning_paper_implementations

🧑‍🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

  • Updated Nov 11, 2025
  • Python

A complete implementation of the "Attention Is All You Need" Transformer model from scratch using PyTorch. This project focuses on building and training a Transformer for neural machine translation (English-to-Italian) on the OpusBooks dataset.

  • Updated Nov 8, 2025
  • Python

Linear-time sequence modeling that replaces attention's O(n²d) complexity with O(nd) summation-based aggregation. Demonstrates constraint-driven emergence: how functional representations can develop from optimization pressure and architectural constraints alone, without explicit pairwise interactions.

  • Updated Nov 3, 2025
  • Python

Improve this page

Add a description, image, and links to the attention topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the attention topic, visit your repo's landing page and select "manage topics."

Learn more