Skip to Main content Skip to Navigation
Conference papers

Modeling the Complexity of Manual Annotation Tasks: a Grid of Analysis

Karen Fort 1, 2 Adeline Nazarenko 2 Sophie Rosset 3
1 SEMAGRAMME - Semantic Analysis of Natural Language
Inria Nancy - Grand Est, LORIA - NLPKD - Department of Natural Language Processing & Knowledge Discovery
LIPN - Laboratoire d'Informatique de Paris-Nord
Abstract : Manual corpus annotation is getting widely used in Natural Language Processing (NLP). While being recognized as a difficult task, no in-depth analysis of its complexity has been performed yet. We provide in this article a grid of analysis of the different complexity dimensions of an annotation task, which helps estimating beforehand the difficulties and cost of annotation campaigns. We observe the applicability of this grid on existing annotation campaigns and detail its application on a real-world example.
Complete list of metadatas

Cited literature [24 references]  Display  Hide  Download
Contributor : Karën Fort <>
Submitted on : Wednesday, January 2, 2013 - 4:37:59 PM
Last modification on : Saturday, February 15, 2020 - 2:02:44 AM
Document(s) archivé(s) le : Friday, March 31, 2017 - 10:19:36 PM


Files produced by the author(s)


  • HAL Id : hal-00769631, version 1


Karen Fort, Adeline Nazarenko, Sophie Rosset. Modeling the Complexity of Manual Annotation Tasks: a Grid of Analysis. International Conference on Computational Linguistics, Dec 2012, Mumbaï, India. pp.895--910. ⟨hal-00769631⟩



Record views


Files downloads