Example: bankruptcy

Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary ...

Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary Cross- modal Relation Detection Xincheng Ju1 , Dong Zhang1 , Rong Xiao2 , Junhui Li1 , Shoushan Li1 , Min Zhang 1 , Guodong Zhou1. 1. School of Computer Science and Technology, Soochow University, China 2. Alibaba Group, China {dzhang, jhli, lishoushan, minzhang, Abstract Aspect terms extraction (ATE) and aspect sen- timent classification (ASC) are two fundamen- tal and fine-grained sub-tasks in aspect-level sentiment Analysis (ALSA). In the textual anal- ysis, jointly extracting both aspect terms and sentiment polarities has been drawn much at- tention due to the better applications than in- [NBA]Neu: [Spurs]Pos rout RT @ funnytwittingg : [OBAMA]Neg TO.}

Multi-modal aspect-level (aka target-oriented) sen-timent analysis (MALSA) is an important and fine-grained task in multi-modal sentiment analysis ... to the details (e.g., different facial expressions) of some regions, so that we can judge the accurate sen-timent dependent on a specific aspect “OBAMA”. Therefore, a well-behaved approach ...

Tags:

  Expression, Modal

Information

Domain:

Source:

Link to this page:

Please notify us if you found a problem with this document:

Other abuse

Transcription of Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary ...

1 Joint Multi-modal Aspect-Sentiment Analysis with Auxiliary Cross- modal Relation Detection Xincheng Ju1 , Dong Zhang1 , Rong Xiao2 , Junhui Li1 , Shoushan Li1 , Min Zhang 1 , Guodong Zhou1. 1. School of Computer Science and Technology, Soochow University, China 2. Alibaba Group, China {dzhang, jhli, lishoushan, minzhang, Abstract Aspect terms extraction (ATE) and aspect sen- timent classification (ASC) are two fundamen- tal and fine-grained sub-tasks in aspect-level sentiment Analysis (ALSA). In the textual anal- ysis, jointly extracting both aspect terms and sentiment polarities has been drawn much at- tention due to the better applications than in- [NBA]Neu: [Spurs]Pos rout RT @ funnytwittingg : [OBAMA]Neg TO.}

2 [ISRAEL]Neu ? [OBAMA]Neg TO. dividual sub-task. However, in the multi- [Thunder]Neg | Tempo Sports [UKRAINE] ? [OBAMA]Neg TO [USA]Neu ? modal scenario, the existing studies are limited (a) (b). to handle each sub-task independently, which fails to model the innate connection between Figure 1: Two examples for Joint Multi-modal aspect- the above two objectives and ignores the better sentiment Analysis . applications. Therefore, in this paper, we are the first to jointly perform Multi-modal ATE. (MATE) and Multi-modal ASC (MASC), and aspect terms from a free text with its accompany- we propose a Multi-modal Joint learning ap- ing image (Wu et al.)

3 , 2020a). Second, MASC aims proach with Auxiliary cross- modal relation de- to classify the sentiment polarity of a Multi-modal tection for Multi-modal aspect-level sentiment Analysis (MALSA). Specifically, we first build post towards a given aspect in textual modality (Yu an Auxiliary text-image relation detection mod- and Jiang, 2019). ule to control the proper exploitation of visual To better satisfy the practical applications, the information. Second, we adopt the hierarchi- aspect term-polarity co-extraction, which solves cal framework to bridge the Multi-modal con- ATE and ASC simultaneously, receives much atten- nection between MATE and MASC, as well tion recently in a textual scenario (Wan et al.

4 , 2020;. as separately visual guiding for each sub mod- Chen and Qian, 2020b; Ying et al., 2020). How- ule. Finally, we can obtain all aspect-level sentiment polarities dependent on the jointly ever, to our best knowledge, in the Multi-modal extracted specific aspects. Extensive experi- scenario, the Joint MATE and MASC, , Joint ments show the effectiveness of our approach Multi-modal Aspect-Sentiment Analysis (JMASA), against the Joint textual approaches, pipeline have never been investigated so far. For this Joint and collapsed Multi-modal approaches. Multi-modal task, we believe that there exist the following challenges at least.

5 1 Introduction On the one hand, visual modality may provide no clues for one of sub-tasks. For example, in Fig- Multi-modal aspect-level (aka target-oriented) sen- ure 1(a), since the image shows most of the content timent Analysis (MALSA) is an important and fine- described in the text, and we can't infer from the grained task in Multi-modal sentiment Analysis image which team has an advantage at first glance. (MSA). Previous studies normally cast MALSA in While, a direct understanding of the text ( , the social media as two independent sub-tasks: Multi- word rout ) seems to be able to judge the senti- modal Aspect Terms Extraction (MATE) and Multi- ment of Spurs and Thunder.

6 Thus this image modal Aspect Sentiment Classification (MASC). does not add to the text tweet meaning (Vempala First, MATE aims to detect a set of all potential and Preotiuc-Pietro, 2019). On the contrary, in . Corresponding Author Figure 1(b), the information of textual modality is 4395. Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 4395 4405. November 7 11, 2021. c 2021 Association for Computational Linguistics quite limited so that we cannot directly infer the need to be applied to the industry recently (Akhtar sentiment towards one aspect. While, the visual et al., 2019; Zadeh et al.)

7 , 2020; Sun et al., 2021a;. modality provides rich clues ( , differential ex- Tang et al., 2019; Zhang et al., 2020b, 2021a). pressions) to help us predict the correct sentiment In the following, we mainly overview the limited of OBAMA . Therefore, a well-behaved approach studies of Multi-modal aspect terms extraction and should determine whether the visual information Multi-modal aspect sentiment classification on text adds to the textual modality (cross- modal relation and image modalities. Besides, we also introduce detection) and how much visual information con- some representative studies for text-based Joint as- tributes to text.

8 Pect terms extraction and sentiment polarity classi- On the other hand, the characteristics of the fication. two Multi-modal sub-tasks are different: one is Multi-modal Aspect Terms Extraction sequence labeling problem, the other is aspect- (MATE). Sequence labeling approaches are dependent classification problem. Different tasks typically employed for this sub-task(Ma et al., seem to focus on different image information. For 2019; Chen and Qian, 2020a; Karamanolakis et al., example, in Figure 1(b), towards first sub-task 2019). But it is challenging to bridge the gap MATE, if we can attend to some coarse-grained between text and image.

9 Several related studies concepts ( , silhouette of human face, Person la- with focus on named entity recognition propose to bel) in the image, it is enough and effective to help leverage the whole image information by ResNet identify the name OBAMA in the text as an aspect. encoding to augment each word representation, Towards second sub-task MASC, we should attend such as (Moon et al., 2018; Zhang et al., 2018). to the details ( , different facial expressions) of upon RNN, (Yu et al., 2020b) upon Transformer some regions, so that we can judge the accurate sen- and (Zhang et al., 2021b) on GNN. Besides, timent dependent on a specific aspect OBAMA.

10 Several related studies propose to leveraging the Therefore, a well-behaved approach should sepa- fine-grained visual information by object detection, rately mine the visual information for these two such as (Wu et al., 2020a,b). sub-tasks instead of collapsed tagging with the However, all the above studies completely ig- same visual feeding. nore the sentiment polarity Analysis dependent on To handle the above challenges, we propose a the detected target, which has great facilitates in Multi-modal Joint learning approach with auxil- practical applications, such as e-commerce. Dif- iary cross- modal relation detection, namely JML.


Related search queries