<会議発表論文>
Improving Interpretability in Document-Level Polarity Classification by Applying Attention

作成者
本文言語
出版者
発行日
収録物名
開始ページ
終了ページ
会議情報
出版タイプ
アクセス権
権利関係
関連DOI
関連DOI
関連ISBN
関連HDL
関連情報
概要 Document-level polarity classification has attracted interest in the real world. While LLMs have made it possible for accurate classification, these complex models have the problem of interpretability.... Our contribution is to apply inter-sentence attention, which captures the relationship between sentences, to a more practical interpretable model. By utilizing high inter-sentence attention scores, meaning corresponding sentences are related to each other, we attempt to capture the context of sentences and make them more similar to the human judgment process. With two real datasets, we compared our model with prior models in terms of classification performance and interpretability and found that our model is more accurate on both datasets. In addition, to assess interpretability, we examined the overlap between sentences that contribute to the model's predictions and those annotated by humans for the same document. The results show that our model has a larger overlap and is more likely to extract interpretive sentences that humans intuitively consider important. In addition, our result partially captures the polarity of “implicit” sentences that do not contain direct expressions, which could not be captured by prior models, suggesting that our model may lead to a more natural interpretation.続きを見る

本文ファイル

pdf 7378126 pdf 213 KB 4  

詳細

PISSN
レコードID
関連ISBN
主題
登録日 2025.08.25
更新日 2025.08.29