Human Emotion Understanding using multimodal dataset.
-
Updated
Jul 27, 2020 - Jupyter Notebook
Human Emotion Understanding using multimodal dataset.
The code for our INTERSPEECH 2020 paper - Jointly Fine-Tuning "BERT-like'" Self Supervised Models to Improve Multimodal Speech Emotion Recognition
Offical implementation of paper "MSAF: Multimodal Split Attention Fusion"
The code for our IEEE ACCESS (2020) paper Multimodal Emotion Recognition with Transformer-Based Self Supervised Feature Fusion.
A survey of deep multimodal emotion recognition.
A Tensorflow implementation of Speech Emotion Recognition using Audio signals and Text Data
All experiments were done to classify multimodal data.
Emotion recognition from Speech & Text using different heterogeneous ensemble learning methods
audio-text multimodal emotion recognition model which is robust to missing data
The repo contains an audio emotion detection model, facial emotion detection model, and a model that combines both these models to predict emotions from a video
Published in Springer Multimedia Tools and Applications Journal.
Lightweight and Interpretable ML Model for Speech Emotion Recognition and Ambiguity Resolution (trained on IEMOCAP dataset)
SERVER: Multi-modal Speech Emotion Recognition using Transformer-based and Vision-based Embeddings
MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversation
This API utilizes a pre-trained model for emotion recognition from audio files. It accepts audio files as input, processes them using the pre-trained model, and returns the predicted emotion along with the confidence score. The API leverages the FastAPI framework for easy development and deployment.
This repository provides the codes for MMA-DFER: multimodal (audiovisual) emotion recognition method. This is an official implementation for the paper MMA-DFER: MultiModal Adaptation of unimodal models for Dynamic Facial Expression Recognition in-the-wild.
A collection of datasets for the purpose of emotion recognition/detection in speech.
Scientific Reports - Open access - Published: 14 February 2025
A Unimodal Valence-Arousal Driven Contrastive Learning Framework for Multimodal Multi-Label Emotion Recognition (ACM MM 2024 oral)
😎 Awesome lists about Speech Emotion Recognition
Add a description, image, and links to the multimodal-emotion-recognition topic page so that developers can more easily learn about it.
To associate your repository with the multimodal-emotion-recognition topic, visit your repo's landing page and select "manage topics."