🎓Automatically Update CV Papers Daily using Github Actions (Update Every 12th hours)
-
Updated
Nov 8, 2025 - Python
🎓Automatically Update CV Papers Daily using Github Actions (Update Every 12th hours)
autoupdate paper list
Train, Evaluate, Optimize, Deploy Computer Vision Models via OpenVINO™
The collection of pre-trained, state-of-the-art AI models for ailia SDK
Code to reproduce a zero-shot action recognition approach on the Epic-Kitchens dataset.
A physics-based video search engine using Meta's V-JEPA 2 world model to find videos with similar motion dynamics.
Computer vision that understands temporal relationships and causality in video sequences.
PyTorch implementation of a Relational Attention-Based Model for Group Activity Recognition — inspired by "A Hierarchical Deep Temporal Model for Group Activity Recognition (arXiv:1511.06040v2)"
[CCS'24] Official Implementation of "Fisher Information guided Purification against Backdoor Attacks"
PySlowFast, the official video understanding framework from Facebook AI Research (FAIR), to train, evaluate, and reproduce state-of-the-art video models on the UCF24 action detection dataset. It supports customizable training pipelines, model fine-tuning, and evaluation for video-based action recognition and spatio-temporal localization tasks.
DLC2Action is an action segmentation package that makes running and tracking of machine learning experiments easy.
Curated catalog of 40+ Human Activity Recognition (HAR) and action recognition datasets: vision, wearable sensors, skeleton/mocap, multimodal. Includes benchmarks, baselines, and citations.
PyTorch implementation of "BDC-CLIP: Brownian Distance Covariance for Adapting CLIP to Action Recognition" (ICML 2025).
A sport-tailored, pose-enhanced action recognition framework
This is the offical repository of LLAVIDAL
A real-time inferencing of multistreaming YOWOv3(Spatio Temporal Action Detection task) using (UCF101-24) dataset. The repo is extension of https://github.com/Hope1337/YOWOv3, https://arxiv.org/pdf/2408.02623
Laparoscopic video dataset for surgical action triplet recognition
A transformer-inspired neural network for surgical action triplet recognition from laparoscopic videos.
Add a description, image, and links to the action-recognition topic page so that developers can more easily learn about it.
To associate your repository with the action-recognition topic, visit your repo's landing page and select "manage topics."