Argument Search
Assessing Argument Relevance
- verfasst von
- Martin Potthast, Lukas Gienapp, Florian Euchner, Nick Heilenkötter, Nico Weidmann, Henning Wachsmuth, Benno Stein, Matthias Hagen
- Abstract
We report on the first user study on assessing argument relevance. Based on a search among more than 300,000 arguments, four standard retrieval models are compared on 40 topics for 20 controversial issues: every issue has one topic with a biased stance and another neutral one. Following TREC, the top results of the different models on a topic were pooled and relevance-judged by one assessor per topic. The assessors also judged the arguments' rhetorical, logical, and dialectical quality, the results of which were cross-referenced with the relevance judgments. Furthermore, the assessors were asked for their personal opinion, and whether it matched the predefined stance of a topic. Among other results, we find that Terrier's implementations of DirichletLM and DPH are on par, significantly outperforming TFIDF and BM25. The judgments of relevance and quality hardly correlate, giving rise to a more diverse set of ranking criteria than relevance alone. We did not measure a significant bias of assessors when their stance is at odds with a topic's stance.
- Externe Organisation(en)
-
Universität Leipzig
Universität Stuttgart
Universität Bremen
Karlsruher Institut für Technologie (KIT)
Universität Paderborn
Bauhaus-Universität Weimar
Martin-Luther-Universität Halle-Wittenberg
- Typ
- Aufsatz in Konferenzband
- Seiten
- 1117-1120
- Anzahl der Seiten
- 4
- Publikationsdatum
- 18.07.2019
- Publikationsstatus
- Veröffentlicht
- Peer-reviewed
- Ja
- ASJC Scopus Sachgebiete
- Information systems, Angewandte Mathematik, Software
- Elektronische Version(en)
-
https://doi.org/10.1145/3331184.3331327 (Zugang:
Geschlossen)