Giter Club home page Giter Club logo

multimodal-sentiment-analysis-paper-list's Introduction

Multimodal-Sentiment-Analysis(MSA) paper list

From September 27, 2019, all future readings will be recorded in the document. Because I will graduate soon, there are many things to do, so the update is slow. After a period of time, I will organize all papers in a large scale during my graduate study

Multimodal Sentiment analysis

1、ICON: Interactive Conversational Memory Network for Multimodal Emotion Detection https://www.aclweb.org/anthology/D18-1280

2、Context-Dependent Sentiment Analysis in User-Generated Videos (ACL 2017).

3、Multi-level Multiple Attentions for Contextual Multimodal Sentiment Analysis(ICDM 2017).

Code for 2、3:https://github.com/SenticNet/multimodal-fusion

4、Tensor Fusion Network for Multimodal Sentiment Analysis(EMNLP 2017)

Code for 4:https://github.com/Justin1904/TensorFusionNetworks

5、Multimodal Transformer for Unaligned Multimodal Language Sequences(ACL 2019)

Code for 5:https://github.com/yaohungt/Multimodal-Transformer

6、Memory Fusion Network for Multi-view Sequential Learning(AAAI 2018)

Code for 6:https://github.com/pliang279/MFN

7、FACTORIZED MULTIMODAL TRANSFORMER FOR MULTIMODAL SEQUENTIAL LEARNING

Code for 7:https://github.com/A2Zadeh/Factorized-Multimodal-Transformer(release April 15th, 2020)

8、Multimodal Transformer for Unaligned Multimodal Language Sequences(ACL 2019)

Code for 8:https://github.com/yaohungt/Multimodal-Transformer

9、Multimodal Sentiment Analysis using Hierarchical Fusion with Context Modeling

Code for 9:https://github.com/SenticNet/hfusion

10、Found in Translation: Learning Robust Joint Representations by Cyclic Translations Between Modalities(2019 AAAI)

Code for 10:https://github.com/hainow/MCTN

11、Words Can Shift:Dynamically AdjustingWord Representations Using Nonverbal Behaviors(2019 AAAI)

Code for 11:https://github.com/victorywys/RAVEN

12、A Transformer-based joint-encoding for Emotion Recognition and Sentiment Analysis(2020 ACL workshop)

Code for 12:https://github.com/jbdel/MOSEI_UMONS

13、Low Rank Fusion based Transformers for Multimodal Sequences(2020 ACL workshop)

14、Quantum Cognitively Motivated Decision Fusion for Video Sentiment Analysis(AAAI2021)

15、Learning Modality-Specific Representations with Self-Supervised Multi-Task Learning for Multimodal Sentiment Analysis(AAAI2021)

16、Quantum Cognitively Motivated Decision Fusion for Video Sentiment Analysis(AAAI2021)

17、An Entanglement-driven Fusion Neural Network for Video Sentiment Analysis(IJCAI2021)

18、Quantum-inspired Neural Network for Conversational Emotion Recognition(AAAI2021)

Multimodal BERT

1、VideoBERT: A Joint Model for Video and Language Representation Learning (ICCV2019)

2、ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks (NeurIPS2019)

3、VisualBERT: A Simple and Performant Baseline for Vision and Language

4、Selfie: Self-supervised Pretraining for Image Embedding

5、Contrastive Bidirectional Transformer for Temporal Representation Learning

6、M-BERT: Injecting Multimodal Information in the BERT Structure

7、LXMERT: Learning Cross-Modality Encoder Representations from Transformers (EMNLP2019)

8、Fusion of Detected Objects in Text for Visual Question Answering (EMNLP2019)

9、Unified Vision-Language Pre-Training for Image Captioning and VQA

Code for 9:https://github.com/LuoweiZhou/VLP

10、VL-BERT: Pre-training of Generic Visual-Linguistic Representations

11、Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training

12、UNITER: Learning UNiversal Image-TExt Representations

13、SpeechBERT: Cross-Modal Pre-trained Language Model for End-to-end Spoken Question Answering

14、Multimodal Transformer for Unaligned Multimodal Language Sequences

Code for 14:https://github.com/yaohungt/Multimodal-Transformer

15、Integrating Multimodal Information in Large Pretrained Transformers(2020 ACL)

Code for 15:https://github.com/WasifurRahman/BERT_multimodal_transformer

16、CM-BERT: Cross-Modal BERT for Text-Audio Sentiment Analysis(2020 ACMMM,ours)

Code for 16:https://github.com/thuiar/Cross-Modal-BERT

Multi-task Sentiment analysis

1、Attention-augmented end-to-end multi-task learning for emotion prediction from speech.

https://arxiv.org/pdf/1903.12424.pdf

2、Multi-task Learning for Multi-modal Emotion Recognition and Sentiment Analysis.

https://arxiv.org/pdf/1905.05812.pdf

3、Multi-task Learning for Target-dependent Sentiment Classification.

https://arxiv.org/pdf/1902.02930.pdf

4、Sentiment and Sarcasm Classification with Multitask Learning.

https://sentic.net/sentiment-and-sarcasm-classification-with-multitask-learning.pdf

Missing Modality

1、SMIL: Multimodal Learning with Severely Missing Modality(2021 AAAI)

Code for 1:https://github.com/mengmenm/SMIL

multimodal-sentiment-analysis-paper-list's People

Contributors

kaicheng-yang0828 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

multimodal-sentiment-analysis-paper-list's Issues

I want to know want is FACET 4.1 in extract vision feature from MOSI

When I extracted the MOSI data set, I read the conversation between you and the author, but I still could not extract the visual mode whose dimension is (batch_size,50,20), I hope you can tell me how to deal with it. I am very curious about the software FACET4.1, can you tell me?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.