-
NanJing Medical University
- Nanjing, Jiangsu Province
Lists (13)
Sort Name ascending (A-Z)
Stars
[NeurIPS 2021] Multiscale Benchmarks for Multimodal Representation Learning
[ACL'19] [PyTorch] Multimodal Transformer
This is the official code repository for "MedMamba: Vision Mamba for Medical Image Classification"
This repository provides implementation for the paper "Self-attention fusion for audiovisual emotion recognition with incomplete data".
Codebase for EMNLP 2024 Findings Paper "Knowledge-Guided Dynamic Modality Attention Fusion Framework for Multimodal Sentiment Analysis"
A curated list of audio-visual learning methods and datasets.
The repo for "Balanced Multimodal Learning via On-the-fly Gradient Modulation", CVPR 2022 (ORAL)
Effective Video Augmentation Techniques for Training Convolutional Neural Networks
Implementation of brand new video augmentation strategy for video action recognition with 3D CNN
The source code for my course completion work (A multimodal emotion recognition system) ๐
Human Emotion Understanding using multimodal dataset.
Emotion recognition with IEMOCAP datasets. We compare the results with SpecAugmentation and CodecAugmentation. For audio codec implementation, we have selected opus.
TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale.
A real time Multimodal Emotion Recognition web app for text, sound and video inputs
EEG Transformer 2.0. i. Convolutional Transformer for EEG Decoding. ii. Novel visualization - Class Activation Topography.
[TAFFC 2024] The official implementation of paper: From Static to Dynamic: Adapting Landmark-Aware Image Models for Facial Expression Recognition in Videos
A Large-scale, Multi-modal, Compound Affective Database for Dynamic Facial Expression Recognition in the Wild.
MAE-DFER: Efficient Masked Autoencoder for Self-supervised Dynamic Facial Expression Recognition (ACM MM 2023)
This repository provides the codes for MMA-DFER: multimodal (audiovisual) emotion recognition method. This is an official implementation for the paper MMA-DFER: MultiModal Adaptation of unimodal moโฆ
Deep learning software to decode EEG, ECG or MEG signals
ABAW3 (CVPRW): A Joint Cross-Attention Model for Audio-Visual Fusion in Dimensional Emotion Recognition
A list of papers on Generative Adversarial (Neural) Networks
A curated list of awesome self-supervised methods