Manual Annotation Task: Difference between revisions

From GM-RKB
Jump to navigation Jump to search
No edit summary
No edit summary
Line 10: Line 10:
** ...
** ...
* <B>Example(s):</B>
* <B>Example(s):</B>
** a [[Text Annotation Task]] that involves marking parts of speech in sentences for linguistic analysis.
** [[Physical Manual Annotation Task]]s, such as:</s>
** a [[Manual Corpus Annotation Task]], such as [[manual text item classification]].
*** a [[Manual Museum Artifact Annotation Task]] where [[museum curator]]s label [[artifact]]s in a [[museum]] with historical and contextual information.</s>
** a [[Manual Chatbot Answer Scoring Task]].
*** a [[Manual Archaeological Annotation Task]] where [[archaeologist]]s tag items in an [[archaeological dig]] with [[metadata]] about their origin and significance.</s>
** a [[Manual Image Annotation Task]], like labeling objects in images for [[Computer Vision]] datasets.
*** a [[Manual Biological Specimen Annotation Task]] where [[lab technician]]s annotate [[biological specimen]]s with [[taxonomic]] and [[collection data]].</s>
** a [[Manual Video Annotation Task]], where videos are annotated for events or object tracking.
*** a [[Manual Geological Sample Annotation Task]] where [[geologist]]s mark [[geological sample]]s with information on their composition and location of discovery.</s>
** a [[Manual Audio Annotation Task]], such as transcribing speech or annotating audio files for sound events.
*** ...
** [[Digital Manual Annotation Task]]s, such as:</s>
*** a [[Manual Document Annotation Task]] where [[data analyst]]s label [[document]]s with [[topic]]s, [[keyword]]s, or other relevant [[metadata]].</s>
*** a [[Manual Image Annotation Task]] where [[image analyst]]s label objects within [[image]]s for use in [[Computer Vision]] [[dataset]]s.</s>
*** a [[Manual Video Annotation Task]] where [[video analyst]]s annotate [[video]]s for events, activities, or object tracking.</s>
*** a [[Manual Audio Annotation Task]] where [[audio transcriber]]s transcribe [[speech]] or annotate [[audio]] for [[sound event]]s.</s>
*** a [[Manual Text Annotation Task]] that involves marking parts of speech in sentences for linguistic analysis.
**** a [[Manual Corpus Annotation Task]], such as [[manual text item classification]].
*** a [[Manual Chatbot Answer Annotation Task]], ...
** ...
** ...
* <B>Counter-Example(s):</B>
* <B>Counter-Example(s):</B>

Revision as of 01:26, 24 June 2024

A Manual Annotation Task is an annotation task that is done by a human annotator.



References

2014

  • (Sabou et al., 2014) ⇒ Marta Sabou, Kalina Bontcheva, Leon Derczynski, and Arno Scharl. (2014). “Corpus Annotation through Crowdsourcing: Towards Best Practice Guidelines.” In: Proc. LREC.
    • QUOTE: Crowdsourcing is an emerging collaborative approach that can be used for the acquisition of annotated corpora and a wide range of other linguistic resources. Although the use of this approach is intensifying in all its key genres (paid-for crowdsourcing, games with a purpose, volunteering-based approaches), the community still lacks a set of best-practice guidelines similar to the annotation best practices for traditional, expert-based corpus acquisition. In this paper we focus on the use of crowdsourcing methods for corpus acquisition and propose a set of best practice guidelines based in our own experiences in this area and an overview of related literature. We also introduce GATE Crowd, a plugin of the GATE platform that relies on these guidelines and offers tool support for using crowdsourcing in a more principled and efficient manner.

      Over the past ten years, Natural Language Processing (NLP) research has been driven forward by a growing volume of annotated corpora, produced by evaluation initiatives such as ACE (ACE, 2004), TAC,[1] SemEval and Senseval, [2] and large annotation projects such as OntoNotes (Hovy et al., 2006). These corpora have been essential for training and domain adaptation of NLP algorithms and their quantitative evaluation, as well as for enabling algorithm comparison and repeatable experimentation. Thanks to these efforts, there are now well-understood best practices in how to create annotations of consistently high quality, by employing, training, and managing groups of linguistic and/or domain experts. This process is referred to as “the science of annotation” (Hovy, 2010).

      More recently, the emergence of crowdsourcing platforms (e.g. paid-for marketplaces such as Amazon Mechanical Turk (AMT) and CrowdFlower (CF); games with a purpose; and volunteer-based platforms such as crowdcrafting), coupled with growth in internet connectivity, motivated NLP researchers to experiment with crowdsourcing as a novel, collaborative approach for obtaining linguistically annotated corpora. The advantages of crowdsourcing over expert-based annotation have already been discussed elsewhere (Fort et al., 2011; Wang et al., 2012), but in a nutshell, crowdsourcing tends to be cheaper and faster. ...

2009


  1. 1 www.nist.gov/tac
  2. www.senseval.org