Publication Details

Argument Search

Assessing Argument Relevance

authored by
Martin Potthast, Lukas Gienapp, Florian Euchner, Nick Heilenkötter, Nico Weidmann, Henning Wachsmuth, Benno Stein, Matthias Hagen
Abstract

We report on the first user study on assessing argument relevance. Based on a search among more than 300,000 arguments, four standard retrieval models are compared on 40 topics for 20 controversial issues: every issue has one topic with a biased stance and another neutral one. Following TREC, the top results of the different models on a topic were pooled and relevance-judged by one assessor per topic. The assessors also judged the arguments' rhetorical, logical, and dialectical quality, the results of which were cross-referenced with the relevance judgments. Furthermore, the assessors were asked for their personal opinion, and whether it matched the predefined stance of a topic. Among other results, we find that Terrier's implementations of DirichletLM and DPH are on par, significantly outperforming TFIDF and BM25. The judgments of relevance and quality hardly correlate, giving rise to a more diverse set of ranking criteria than relevance alone. We did not measure a significant bias of assessors when their stance is at odds with a topic's stance.

External Organisation(s)
Leipzig University
University of Stuttgart
University of Bremen
Karlsruhe Institute of Technology (KIT)
Paderborn University
Bauhaus-Universität Weimar
Martin Luther University Halle-Wittenberg
Type
Conference contribution
Pages
1117-1120
No. of pages
4
Publication date
18.07.2019
Publication status
Published
Peer reviewed
Yes
ASJC Scopus subject areas
Information Systems, Applied Mathematics, Software
Electronic version(s)
https://doi.org/10.1145/3331184.3331327 (Access: Closed)