Publication Details

ArgInstruct: Specialized Instruction Fine-Tuning for Computational Argumentation

Authored by

Maja Stahl, Timon Ziegenbein, Joonsuk Park, Henning Wachsmuth

Abstract

Training large language models (LLMs) to follow instructions has significantly enhanced their ability to tackle unseen tasks. However, despite their strong generalization capabilities, instruction-following LLMs encounter difficulties when dealing with tasks that require domain knowledge. This work introduces a specialized instruction fine-tuning for the domain of computational argumentation (CA). The goal is to enable an LLM to effectively tackle any unseen CA tasks while preserving its generalization capabilities. Reviewing existing CA research, we crafted natural language instructions for 105 CA tasks to this end. On this basis, we developed a CA-specific benchmark for LLMs that allows for a comprehensive evaluation of LLMs' capabilities in solving various CA tasks. We synthesized 52k CA-related instructions, adapting the self-instruct process to train a CA-specialized instruction-following LLM. Our experiments suggest that CA-specialized instruction fine-tuning significantly enhances the LLM on both seen and unseen CA tasks. At the same time, performance on the general NLP tasks of the SuperNI benchmark remains stable.

Details

Organisation(s)
Natural Language Processing Section
Type
Conference contribution
Pages
11103–11127
No. of pages
25
Publication date
27.07.2025
Publication status
Published
Peer reviewed
Yes
ASJC Scopus subject areas
Language and Linguistics, Linguistics and Language, Computer Science Applications
Electronic version(s)
https://doi.org/10.18653/v1/2025.findings-acl.579 (Access: Open )

Cite

Loading...