直近一年間の累計
アクセス数 : ?
ダウンロード数 : ?
ID 118516
著者
Ding, Fei Tokushima University
任, 福継 University of Electronic Science and Technology of China 徳島大学 教育研究者総覧 KAKEN研究者をさがす
キーワード
Neuro-symbolic AI
Sentiment Analysis
Fine-tuned Transformer
Latent Dirichlet Allocation
資料タイプ
学術雑誌論文
抄録
For text sentiment analysis, state-of-the-art neural language models have demonstrated promising performance. However, they lack interpretability, require vast volumes of annotated data, and are typically specialized for tasks. In this paper, we explore a connection between fine-tuned Transformer models and unsupervised LDA approach to cope with text sentiment analysis tasks, inspired by the concept of Neuro-symbolic AI. The Transformer and LDA models are combined as a feature extractor to extract the hidden representations of the input text sequences. Subsequently, we employ a feedforward network to forecast various sentiment analysis tasks, such as multi-label emotion prediction, dialogue quality prediction, and nugget detection. Our proposed method obtains the best results in the NTCIR-16 dialogue evaluation (DialEval-2) task, as well as cutting-edge results in emotional intensity prediction using the Ren_CECps corpus. Extensive experiments show that our proposed method is highly explainable, cost-effective in training, and superior in terms of accuracy and robustness.
掲載誌名
IEEE Transactions on Affective Computing
ISSN
19493045
出版者
IEEE
15
2
開始ページ
493
終了ページ
507
発行日
2023-05-23
備考
論文本文は2025-05-23以降公開予定
権利情報
© 2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
EDB ID
出版社版DOI
出版社版URL
言語
eng
著者版フラグ
その他
部局
理工学系