Assessing the Quality of Scientific Output in its Ecosystem

Motivation and Objectives

This is the next iteration of the successful Semantic Publishing Challenge of ESWC 2014 and 2015. We continue pursuing the objective of assessing the quality of scientific output, evolving the dataset bootstrapped in 2014 and 2015 to take into account the wider ecosystem of publications. To achieve that, this year’s challenge focuses on refining and enriching an existing linked open dataset about workshops, their publications and their authors. Aspects of “refining and enriching” include extracting deeper information from the HTML and PDF sources of the workshop proceedings volumes and enriching this information with knowledge from existing datasets. Thus, a combination of broadly investigated technologies in the Semantic Web field, such as Information Extraction (IE), Natural Language Processing (NLP), Named Entity Recognition (NER), link discovery, etc., is required to deal with the challenge’s tasks.

Target Audience

The Challenge is open to everyone from industry and academia

Task

We ask challengers to automatically annotate a set of multi-format input documents and to produce a LOD that fully describes these documents, their context, and relevant parts of their content. The evaluation will consist of evaluating a set of queries against the produced dataset to assess its correctness and completeness. The primary input dataset is the LOD that has been extracted from the CEURWS.org workshop proceedings using the winning extraction tools of the 2014 and 2015 challenges, plus its full original HTML and PDF source documents. In addition, the challenge uses (as linking targets) existing LOD on scholarly publications. The input dataset will be split in two parts: a training dataset and an evaluation dataset, which will disclosed a few days before the submission deadline. Participants will be asked to run their tool on the evaluation dataset and to produce the final Linked Dataset and the output of the queries on that dataset. The Challenge includes three tasks:

 

Task 1: Extraction and assessment of workshop proceedings information in HTML

Participants are required to extract information from a set of HTML tables of contents published in CEUR-WS.org workshop proceedings. The extracted information is expected to answer queries about the quality of these workshops, for instance by measuring growth, longevity, etc. The task is an extension of the Task 1 of the 2015 and 20154 Challenge,: we will reuse the most challenging quality indicators from last year’s challenge, others will bedefined more precisely, others will be completely new. Last years’ results, with an F-measure of 0.66 in 2015 and 0.64 in 2014 for the winning solutions, show improvement but there is a lot of room for ameliorating information extraction.

Task 2: Extracting information from the full text of the papers

Participants are required to extract information from the textual content of the paper, extracted from PDF files that are mentioned in Task 1, as it occurred in 2015 and in contrast to 2014 that XML files from publications ofanother domain were used. That information should provide a deeper understanding of the context in which a paper was written. In particular, the extracted information is expected to answer queries about authors’ affiliations and research institutions, along with research fundings and projects. The task mainly requires NLP and Entity Recognition techniques.

Task 3: Interlinking

Participants are required to interlink the CEUR-WS.org linked dataset with relevant datasets already existing in the LOD cloud. Task 3 can be accomplished as an entity interlinking/instance matching task that aims to address both interlinking data from the output of the other tasks as well as interlinking CEUR-WS.org linked dataset to external datasets. Moreover, as triples are generated from different sources and due to different activities, tracking provenance information becomes increasingly important.

 

Evaluation

Participants will be requested to submit the LOD that their tool produces from the evaluation dataset, as well as a paper that describes their approach. They will also be given a set of queries in natural language form and will be asked to translate those queries into a SPARQL form that works on their LOD.

This year there are two important changes:

  • participants will be provided a tool for evaluating their queries during the training phase. They will also be given a set of CSV that the queries are expected to produce on the training dataset.
  • a few days before the deadline we will publish the set of query used for the final evaluation. Participants will be asked then to run these queries on their dataset and to submit the produced output in CSV

The results of these queries will be compared with the expected output, and precision and recall will be measured to identify the best performing approach. We will use the same evaluation tool made available during the training phase.

Separately, the most original approach will be assigned by the Program Committee.

For details, see our general rules.

 

Feedback and Discussion

A discussion group is open for participants to ask questions and to receive updates about the challenge: mailto: sempub-challenge@googlegroups.com. Participants are invited to subscribe to this group as soon as possible and to communicate their intention to participate. They are also invited to use this channel to discuss problems in the input dataset and to suggest changes.

 

How to Participate

Participants are required to submit:

  • Abstract: no more than 200 words.
  • Description: It should explain the details of the automated annotation system, including why the system is innovative, how it uses Semantic Web technology, what features or functions the system provides, what design choices were made and what lessons were learned. The description should also summarize how participants have addressed the evaluation tasks. An outlook towards how the data could be consumed is appreciated but not strictly required. Papers must be submitted in PDF format, following the style of the Springer's Lecture Notes in Computer Science (LNCS) series, and not exceeding 12 pages in length.
  • The Linked Open Dataset produced by their tool on the evaluation dataset (as a file or as a URL, in Turtle or RDF/XML).
  • A set of SPARQL queries that work on that LOD and correspond to the natural language queries provided as input
  • The output of these SPARQL queries on the evaluation dataset (in CSV format)

Participants will also be asked to submit their tool (source and/or binaries, or a link these can be downloaded from, or a web service URL) for verification purposes. Further submission instructions will be published on the challenge wiki.

All submissions should be provided via the submission system linked from the homepage.

Note: Papers will be included in the Conference USB stick. After the conference, participants will be able to add data about the evaluation and to finalize the camera-ready for the final proceedings.

 

Judging and Prizes

After a first round of review, the Program Committee and the chairs will select a number of submissions conforming to the challenge requirements that will be invited to present their work. Submissions accepted for presentation will receive constructive reviews from the Program Committee, they will be included in the Springer post-proceedings of ESWC. More details will follow.

 

Six winners will be selected. For each task we will select:

  • best performing tool, given to the paper which will get the highest score in the evaluation
  • most original approach, selected by the Challenge Committee with the reviewing process

 

Important Dates

  • Publication of the full description of tasks, rules and queries; publication of the training dataset: January 20th, 2016
  • Publication of the evaluation tool: February 15th, 2016
  • Abstract submission: February 26th, 2016   
  • Paper submission: March 11th, 2016
  • Deadline for making remarks to the training dataset and the evaluation tool: March 31th, 2016
  • Notification and invitation to submit task results: ​April 8th, 2016
  • Conference camera-ready: April 24th, 2016
  • Publication of the evaluation dataset details: May 11th, 2016
  • Results submission: May 13th, 2016 
  • Challenge days: May 29th June 2nd, 2016

Program Committee

  • Aliaksandr Birukou, Springer Verlag, Heidelberg, Germany
  • Lukasz Bolikowski, University of Warsaw, Poland
  • Kai Eckert, University of Mannheim, Germany
  • Maxim Kolchin, ITMO University, SaintPetersburg, Russia
  • Phillip Lord, Newcastle University, UK
  • Philipp Mayr, GESIS, Germany
  • Jodi Schneider, University of Pittsburgh, USA
  • Selver Softic, Graz University of Technology, Austria
  • Ruben Verborgh, Ghent university – iMinds
  • Michael Wagner, Schloss Dagstuhl, LeibnizZentrum für Informatik, German

 

Organizing Committee:

  • Angelo Di Iorio, Department of Computer Science and Engineering, University of Bologna, IT
  • Anastasia Dimou, Ghent University, BE
  • Christoph Lange, Enterprise Information Systems, University of Bonn / Fraunhofer IAIS, DE
  • Sahar Vahdati, University of Bonn, DE

 

Websitehttps://github.com/ceurws/lod/wiki/SemPub2016