Skip to Main content Skip to Navigation
Conference papers

Modeling the Complexity of Manual Annotation Tasks: a Grid of Analysis

Karen Fort 1, 2 Adeline Nazarenko 2 Sophie Rosset 3
1 SEMAGRAMME - Semantic Analysis of Natural Language
Inria Nancy - Grand Est, LORIA - NLPKD - Department of Natural Language Processing & Knowledge Discovery
2 RCLN
LIPN - Laboratoire d'Informatique de Paris-Nord
Abstract : Manual corpus annotation is getting widely used in Natural Language Processing (NLP). While being recognized as a difficult task, no in-depth analysis of its complexity has been performed yet. We provide in this article a grid of analysis of the different complexity dimensions of an annotation task, which helps estimating beforehand the difficulties and cost of annotation campaigns. We observe the applicability of this grid on existing annotation campaigns and detail its application on a real-world example.
Complete list of metadatas

Cited literature [24 references]  Display  Hide  Download

https://hal.archives-ouvertes.fr/hal-00769631
Contributor : Karën Fort <>
Submitted on : Wednesday, January 2, 2013 - 4:37:59 PM
Last modification on : Saturday, February 15, 2020 - 2:02:44 AM
Document(s) archivé(s) le : Friday, March 31, 2017 - 10:19:36 PM

File

coling2012_Complexity_KF_30102...
Files produced by the author(s)

Identifiers

  • HAL Id : hal-00769631, version 1

Citation

Karen Fort, Adeline Nazarenko, Sophie Rosset. Modeling the Complexity of Manual Annotation Tasks: a Grid of Analysis. International Conference on Computational Linguistics, Dec 2012, Mumbaï, India. pp.895--910. ⟨hal-00769631⟩

Share

Metrics

Record views

768

Files downloads

416