-
University of Central Florida
- Orlando, Florida
- www.aritradutta.com
Stars
An SVD-free Proximal Gradient & Alternating Minimization Method for Low-rank Recovery
This code is provided for reproducibility of results in the paper: Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve Aerial Visual Perception?
[WACV 2026 🔥] GAEA is a multimodal model with a new dataset and benchmark for context-aware image geolocation and QA.
Kolmogorov-Arnold Attention: Is Learnable Attention Better for Vision Transformers?
Awesome Reasoning LLM Tutorial/Survey/Guide
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Inceptive Visual Representation Learning with Diverse Attention Across Heads. Image Classification, Action Recognition, and Robot Learning.
Official repository for Towards Multi-modal Transformers in Federated Learning (ECCV2024)
Official Repository of "Fibottention: Inceptive Visual Representation Learning with Diverse Attention Across Heads"
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
aritra-dutta / DeepReduce-1
Forked from hangxu0304/DeepReduceA Sparse-tensor Communication Framework for Distributed Deep Learning
GRACE - GRAdient ComprEssion for distributed deep learning