谁的故事被讲述?大语言模型在人生叙事摘要中的立场性与偏见 / Whose Story Gets Told? Positionality and Bias in LLM Summaries of Life Narratives
1️⃣ 一句话总结
这篇论文研究发现,当使用大语言模型来总结和分析人们的人生故事时,模型会因种族和性别偏见而扭曲叙事视角,可能导致代表性伤害,因此作者建议未来研究应评估模型在解读文本时的立场性。
Increasingly, studies are exploring using Large Language Models (LLMs) for accelerated or scaled qualitative analysis of text data. While we can compare LLM accuracy against human labels directly for deductive coding, or labeling text, it is more challenging to judge the ethics and effectiveness of using LLMs in abstractive methods such as inductive thematic analysis. We collaborate with psychologists to study the abstractive claims LLMs make about human life stories, asking, how does using an LLM as an interpreter of meaning affect the conclusions and perspectives of a study? We propose a summarization-based pipeline for surfacing biases in perspective-taking an LLM might employ in interpreting these life stories. We demonstrate that our pipeline can identify both race and gender bias with the potential for representational harm. Finally, we encourage the use of this analysis in future studies involving LLM-based interpretation of study participants' written text or transcribed speech to characterize a positionality portrait for the study.
谁的故事被讲述?大语言模型在人生叙事摘要中的立场性与偏见 / Whose Story Gets Told? Positionality and Bias in LLM Summaries of Life Narratives
这篇论文研究发现,当使用大语言模型来总结和分析人们的人生故事时,模型会因种族和性别偏见而扭曲叙事视角,可能导致代表性伤害,因此作者建议未来研究应评估模型在解读文本时的立场性。
源自 arXiv: 2604.20131