Home

Awesome

A Survey of NLP Annotation Platforms

This README is a summary of responses to questionnaire on annotation platforms (from https://forms.gle/iZk8kehkjAWmB8xe9). The questionnaire is a short survey on users' usage and wants of annotation for text and/or image processing for Natural Language Processing (NLP) tasks.

This summary is based on the results collated on 30 June 2020. We may update the results if there are significant no. of new responses after the stipulate dated in the previous sentence.

The raw data for the survey results can be found on:

Overview

Population

Population Breakdown

Annotations Requirements

Why do you need annotations for your task/data?

What NLP tasks do you need annotation for?

Would the annotations you need require domain expertise or can it be handled by crowdsource workers with some minimum requirements?

Do you have a pool of trusted/expert annotators that can work on your annotation task(s)?

Annotation Platforms

Have you used open source / commercial annotation platforms before?

Which of these open source annotation platforms/tools have you used before?

Which of these commercial annotation platforms/tools have you heard before?

(Note: Fiverr and Upwork are generally crowd-source sites that provides huamn annotators but may/may not provide an annotation platform)

Which of these commercial annotation platforms/tools have you used before?

Note: The response for this question is really low, so it might not be representative of all annotation platform users. But this also highlights the stark adoption rate between open source vs commercial annotation tools.

Most Useful Features

The table below presents the list of features that respondents find useful in the open source or commercial annotation tools they have previously interacted with.

FeatureOpen SourceCommercial
Active Learning
Annotation progress monitor
Annotation shortcuts
Annotation visualization
Automatically suggesting annotations
Audio annotation support
Available Online (No installation)
Connecting to external resources (e.g. storage / knowledge base / dictionary)
Customizable annotation tasks/labels (with extra code/schema)
Customizable annotation view (when annotating)
Drag and drop interactions
Easy setup/installation (e.g Docker)
Export to multiple formats
Good UI/UX
Interoperability (e.g. load/combine annotations from other tools, integration with other tools)
Multi-annotator agreement mechanisms/metrics / Automatic evaluation of annotations
Post-Annotation curation
Project Management with collaboration features
Python-based
Simple Login/Sign-on
Supports multiple tasks
Communication tools (e.g. annotators interaction with project managers)
Documentation (e.g. forums, example setup)
Built-in quality control (e.g. screening tests, data cleanup/filer)
Demo/test small projects with least setup effort
Access to large/diverse/global pool of annotators

Suggestions to annotation tool creators

The keywords flexibility and ease/easy/simple appears in many of the comments to list the top features for annotation platform. We suggest the following for annotation tools creators to accomodate these feedbacks:

Another recurring theme in the top feature list includes:

Any other feedback on annotation tools?

Here are a couple of aggregated free-form feedbacks from our respondents:

Feedback for Open Source annotation tools:

Feedback for annotation tools in general:

Your Dream Annotation Platform

Summarizing the respondents' dream platforms:

Notable Mentions about "Your Dream Anntoation Platform"

Acknowledgement

We thank all participants/respondents of the questionnaire and the precious insights/feedbacks given!!!

We also have a thankful mention to Mariana Neves for referring us to the following survey

Cite

If reference to this report is necessary,

to cite (in-text):

Liling Tan. 2020. A Survey of NLP Annotation Platforms. Retrieved from https://github.com/alvations/annotate-questionnaire

in bibtex:

@misc{survey-annotation-platform,
author =   {Liling Tan},
title =    {A Survey of NLP Annotation Platforms},
howpublished = {https://github.com/alvations/annotate-questionnaire},
year = {2020}
}