posted by organizer: giusepperizzo || 3566 views || tracked by 12 users: [display]

#Micropost2014 Challenge 2014 : CfP: Entity Extraction and Linking Challenge (#Microposts2014 @ WWW2014)


When Dec 20, 2013 - Feb 14, 2014
Where Seoul
Submission Deadline Feb 21, 2014
Notification Due Mar 14, 2014
Final Version Due Mar 24, 2014
Categories    NLP   semanticweb   information extraction

Call For Papers

Entity Extraction and Linking Challenge
at the 4th Making Sense of Microposts Workshop
(#Microposts2014) @ WWW 2014
7 April 2014, Seoul, Republic of Korea

Microposts are a highly popular medium to share facts, opinions or emotions. They are an invaluable wealth of data, ready to be mined for training predictive modelings. This year the #Microposts 2014 Workshop will host an "Entity Extraction and Linking Challenge".
The overall task of the challenge is to automatically extract entities from English microposts, and link them to the corresponding English DBpedia v3.9 resources (if the linkage exists). As linking stage we aim to to disambiguate
expressions that are formed by discrete (and typically short) sequences of words.
Existing entity linking tools are intended for use over news corpora and similar document-based corpora with relatively long length. We organise this challenge to foster research into novel, more accurate solutions for the automatic entity linking in (much shorter) micropost data.
We will ask the participants to automatically extract entities (e.g., Obama, London, Rakuten) belonging to all entity types (e.g., Person, Location, Organisation) from a collection of microposts. Participants will have to automatically provide context-relevant DBpedia resources for each entity in a micropost.

The dataset comprises of 3.5K tweets extracted from a much larger collection of over 18 million tweets. This collection, provided by the Redites project (, covers event-annotated tweets collected for the period of 15th July 2011 to 15th August 2011 (31 days). It extends over multiple noteworthy events including the death of Amy Winhehouse, the London Riots and the Oslo bombing. Since the task of this challenge is to automatically extract and link entities, we have built our dataset considering both event and non-event tweets. While event tweets are more likely to contain entities, non-event tweets enable us to evaluate the performance of the system in avoiding false positives in the entity extraction phase.

The dataset has been split into a training (70%) and testing (30%) sets. Following the Twitter TOS we will only provide tweet IDs and annotations for the training set; and tweet IDs for the test set. We will also provide a common framework to mine these datasets from Twitter.

The training set will be released as tsv file where each line consists of :
- tweet_id
- entity_mention_1
- entity_uri_1

- entity_mention_n
- entity_uri_n
Tokens are separated by TABs. Entity mentions and uris are listed according to their appearance order in the tweet.

We will timely advertise the release of the data sets on the workshop mailing list. Please subscribe to More information about dates are available in the Challenge website.

The evaluation consists of two separated stages:
1.- Paper peer review : A community of experts of the domain will judge the quality and applicability of the approaches taken, to provide useful insights on your research;
2.- Precision and Recall: F1 (F-measure with beta = 1) will be computed on a gold standard manually created from the test set. The automatically extracted entities and links will be both matched against this ground truth.

All submissions will be only ranked according to the F1 of each best submission.

Submissions should be provided as a zip file using your system name as the file name (e.g. ''), containing:

1. a TSV file with your system name (e.g. 'awesome.tsv'). We accept up to 3 different submissions, and we will consider *only* the best. If you do so you must specify clearly in your paper the modifications applied to each labelled submission. In this case the submission should contain each of up to 3 TSV files with the tool/system name with "_n" appended to each (e.g. awesome_1.tsv, awesome_2.tsv, awesome_3 ).
In order to evaluate your submissions we require you to submit a tsv file following the format in which the training set is provided.

2. a paper of 6 pages describing your approach and how you tuned/tested it using the training split. All submissions must be in English. Submissions must be in PDF formatted in the style of the Springer Publications format for Lecture Notes in Computer Science (LNCS) []. For details on the LNCS style, see Springer’s Author Instructions. All submissions are not anonymous. Please send us your submission before the deadline through Easychair []. All accepted submissions will be invited for short presentations during the #Microposts2014 workshop and will be published independently from the workshop proceedings on the challenge page and on CEUR [] (note that a minimum number of papers should be submitted in order to be able to publish them on CEUR).

Intent to participate: 13 Jan 2014 (soft)
Release of training set: 14 Jan 2014
Release of test set: 17 Feb 2014
Challenge Submission deadline: 21 Feb 2014 (hard)
Challenge Notification: 14 Mar 2014 (hard)
Challenge camera-ready deadline: 24 Mar 2014 (hard)

Workshop program issued: 15 Mar 2014
Challenge proceedings to be published via CEUR
Workshop - 07 Apr 2014 (Registration open to all)
(All deadlines 23:59 Hawaii Time)

to be announced

Facebook Group:!/home.php?sk=group_180472611974910
Facebook Public Event page:
Google group :!forum/microposts2014
Twitter hashtag: #microposts2014challenge
Twitter account: @Microposts2014
W3C Microposts Community Group:

Challenge Organizers:
Challenge Chair:
A. Elizabeth Cano, Aston University, UK
Giuseppe Rizzo, Università di Torino, Italy

Dataset Chair:
Andrea Varga, The University of Sheffield, UK

Challenge Committee:
Ebrahim Bagheri, Ryerson University, Canada
Pierpaolo Basile, Dipartimento di Informatica - University of Bari, Italy
Uldis Bojars, SIOC Project
Óscar Corcho, Universidad Politécnica de Madrid, Spain
Leon Derczynski, The University of Sheffield, UK
Guillaume Erétéo, Orange Labs
Miriam Fernandez, Knowledge Media Institute, The Open University, UK
Andrés García-Silva, Ontology Engineering Group, Facultad de Informática, Univesidad Politécnica de Madrid, Spain
Anna Lisa Gentile, The University of Sheffield, UK
Robert Jäschke, L3S Research Center, Germany
Diana Maynard, The University of Sheffield, UK
José M. Morales-Del-Castillo, El Colegio de México, Mexico
Georgios Paltoglou, University of Wolverhampton, UK
Bernardo Pereira Nunes, PUC-Rio, Brazil
Daniel Preoţiuc-Pietro, The University of Sheffield, UK
Raphaël Troncy, EURECOM, France
Mischa Tuffield, PeerIndex
Victoria Uren, Aston University, UK

Related Resources

OKE 2016   OKE2016 - Open Knowledge Extraction Challenge @ESWC2016
ACL 2017   The 55th annual meeting of the Association for Computational Linguistics
RuleML Challenge 2016   10th International Rule Challenge 2016
SLIE 2017   Semantic, Logics, Information Extraction and AI
EACL SRW 2017   EACL 2017 Student Research Workshop
11th ODC 2017   11th Oxford Dysfluency Conference - Challenge and Change
RCVC 2016   Robotics & Computer Vision Challenge
NoDaLiDa 2017   21st Nordic Conference on Computational Linguistics
ENSEC 2016   The 2016 ENtity Summarization Evaluation Campaign (ENSEC 2016)
CyS-PDAA 2016   Computación y Sistemas Special Issue on Plagiarism Detection and Authorship Analysis