Authors
Frank Hopfgartner, Allan Hanbury, Henning Müller, Ivan Eggel, Krisztian Balog, Torben Brodt, Gordon V Cormack, Jimmy Lin, Jayashree Kalpathy-Cramer, Noriko Kando, Makoto P Kato, Anastasia Krithara, Tim Gollub, Martin Potthast, Evelyne Viegas, Simon Mercer
Publication date
2018/10/29
Source
Journal of Data and Information Quality (JDIQ)
Volume
10
Issue
4
Pages
1-32
Publisher
ACM
Description
Evaluation in empirical computer science is essential to show progress and assess technologies developed. Several research domains such as information retrieval have long relied on systematic evaluation to measure progress: here, the Cranfield paradigm of creating shared test collections, defining search tasks, and collecting ground truth for these tasks has persisted up until now. In recent years, however, several new challenges have emerged that do not fit this paradigm very well: extremely large data sets, confidential data sets as found in the medical domain, and rapidly changing data sets as often encountered in industry. Crowdsourcing has also changed the way in which industry approaches problem-solving with companies now organizing challenges and handing out monetary awards to incentivize people to work on their challenges, particularly in the field of machine learning.
This article is based on …
Total citations
201920202021202220232024593551
Scholar articles
F Hopfgartner, A Hanbury, H Müller, I Eggel, K Balog… - Journal of Data and Information Quality (JDIQ), 2018