Publication Details

Towards a Perspectivist Turn in Argument Quality Assessment

Authored by

Julia Romberg, Maximilian Maurer, Henning Wachsmuth, Gabriella Lapesa

Abstract

The assessment of argument quality depends on well-established logical, rhetorical, and dialectical properties that are unavoidably subjective: multiple valid assessments may exist, there is no unequivocal ground truth. This aligns with recent paths in machine learning, which embrace the co-existence of different perspectives. However, this potential remains largely unexplored in NLP research on argument quality. One crucial reason seems to be the yet unexplored availability of suitable datasets. We fill this gap by conducting a systematic review of argument quality datasets. We assign them to a multi-layered categorization targeting two aspects: (a) What has been annotated: we collect the quality dimensions covered in datasets and consolidate them in an overarching taxonomy, increasing dataset comparability and interoperability. (b) Who annotated: we survey what information is given about annotators, enabling perspectivist research and grounding our recommendations for future actions. To this end, we discuss datasets suitable for developing perspectivist models (i.e., those containing individual, non-aggregated annotations), and we showcase the importance of a controlled selection of annotators in a pilot study.

Details

Organisation(s)
Natural Language Processing Section
Institute of Artificial Intelligence
External Organisation(s)
GESIS - Leibniz Institute for the Social Sciences
University Hospital Düsseldorf
Type
Conference contribution
Pages
7458-7485
No. of pages
28
Publication date
04.2025
Publication status
Published
Peer reviewed
Yes
ASJC Scopus subject areas
Computer Networks and Communications, Hardware and Architecture, Information Systems, Software
Electronic version(s)
https://doi.org/10.18653/v1/2025.naacl-long.382 (Access: Open )

Cite

Loading...