You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Experiment, data, and analysis code for "Text embedding models yield detailed conceptual knowledge maps derived from short multiple-choice quizzes" by Paxton C. Fitzpatrick, Andrew C. Heusser, and Jeremy R. Manning
This project implements a hierarchical deep learning approach for recognizing group activities in videos. The core idea is to model both individual person actions and group-level activities using temporal dynamics captured by LSTM networks.
llama-2 from scratch — a clean, educational pytorch implementation of the llama-2 transformer architecture. features grouped query attention (gqa), rotary position embeddings (rope), kv caching, and swiglu activation. run easily on google colab and experiment with your own datasets.
attention is all you need — pytorch implementation of the original transformer architecture for english to nepali neural machine translation (nmt), achieving around 27 bleu score.