Skip to main navigation Skip to search Skip to main content

Mitigating Biases of Large Language Models in Stance Detection with Counterfactual Augmented Calibration

  • Ang Li
  • , Jingqian Zhao
  • , Bin Liang
  • , Lin Gui
  • , Hui Wang
  • , Xi Zeng
  • , Xingwei Liang
  • , Kam Fai Wong
  • , Ruifeng Xu*
  • *Corresponding author for this work
  • Harbin Institute of Technology Shenzhen
  • Chinese University of Hong Kong
  • CUHK
  • King's College London
  • Peng Cheng Laboratory
  • China Electronics Technology Group Corporation

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Stance detection is critical for understanding the underlying position or attitude expressed toward a topic. Large language models (LLMs) have demonstrated significant advancements across various natural language processing tasks including stance detection, however, their performance in stance detection is limited by biases and spurious correlations inherent due to their data-driven nature. Our statistical experiment reveals that LLMs are prone to generate biased stances due to sentiment-stance spurious correlations and preference towards certain individuals and topics. Furthermore, the results demonstrate a strong negative correlation between stance bias and stance detection performance, underscoring the importance of mitigating bias to enhance the utility of LLMs in stance detection. Therefore, in this paper, we propose a Counterfactual Augmented Calibration Network (FACTUAL), which a novel calibration network is devised to calibrate potential bias in the stance prediction of LLMs. Further, to address the challenge of effectively learning bias representations and the difficulty in the generalizability of debiasing, we construct counterfactual augmented data. This approach enhances the calibration network, facilitating the debiasing and out-of-domain generalization. Experimental results on in-target and zero-shot stance detection tasks show that the proposed FACTUAL can effectively mitigate biases of LLMs, achieving state-of-the-art results.

Original languageEnglish
Title of host publicationLong Papers
EditorsLuis Chiruzzo, Alan Ritter, Lu Wang
PublisherAssociation for Computational Linguistics (ACL)
Pages7075-7092
Number of pages18
ISBN (Electronic)9798891761896
DOIs
StatePublished - 2025
Externally publishedYes
Event2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2025 - Hybrid, Albuquerque, United States
Duration: 29 Apr 20254 May 2025

Publication series

NameProceedings of the 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies: Long Papers, NAACL-HLT 2025
Volume1

Conference

Conference2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2025
Country/TerritoryUnited States
CityHybrid, Albuquerque
Period29/04/254/05/25

Fingerprint

Dive into the research topics of 'Mitigating Biases of Large Language Models in Stance Detection with Counterfactual Augmented Calibration'. Together they form a unique fingerprint.

Cite this