作成者 |
|
|
|
|
本文言語 |
|
出版者 |
|
発行日 |
|
収録物名 |
|
巻 |
|
号 |
|
開始ページ |
|
会議情報 |
|
出版タイプ |
|
アクセス権 |
|
関連DOI |
|
概要 |
This paper presents results of experiments in which annotators were asked to selectively reexamine their decisions when those decisions seemed inconsistent. The annotation task was binary topic classi...fication. To operationalize the concept of annotationconsistency, a text classifier was trained on all manual annotations made during a complete first pass and then used to automatically recode every document. Annotators were then asked to perform a second manual pass, limiting their attention to cases in which their first annotation disagreed with the text classifier. On average across three annotators, each working independently, 11% of first pass annotations were reconsidered, 46% of reconsidered annotations were changed in the second pass, and 71% of changed annotations agreed with decisions made independently by an experienced fourth annotator. The net result was that for an 11% average increase in annotation cost it was possible to increase overall chance corrected agreement with the annotation decisions of an experienced annotator (as measured by kappa) from 0.70 to 0.75.続きを見る
|