Can LLMs Understand the Impact of Trauma? Costs and Benefits of LLMs Coding the Interviews of Firearm Violence Survivors
Jessica H. Zhu, Shayla Stringfield, Vahe Zaprosyan, Michael Wagner, Michel Cukier + 1 more
TLDR
This study explores using LLMs to code interviews of firearm violence survivors, finding potential but also significant limitations and ethical concerns.
Key contributions
- Assessed open-source LLMs for inductive coding of interviews with 21 Black male firearm violence survivors.
- Found some LLM configurations can identify important qualitative codes.
- Revealed overall code relevance remains low and is highly sensitive to data processing.
- Identified that LLM guardrails cause substantial narrative erasure in sensitive data.
Why it matters
This paper is crucial for understanding the practical and ethical challenges of applying LLMs to sensitive qualitative research with vulnerable populations. It highlights that while AI offers scalability, its current limitations and inherent biases, like guardrails causing narrative loss, demand careful consideration before widespread adoption.
Original Abstract
Firearm violence is a pressing public health issue, yet research into survivors' lived experiences remains underfunded and difficult to scale. Qualitative research, including in-depth interviews, is a valuable tool for understanding the personal and societal consequences of community firearm violence and designing effective interventions. However, manually analyzing these narratives through thematic analysis and inductive coding is time-consuming and labor-intensive. Recent advancements in large language models (LLMs) have opened the door to automating this process, though concerns remain about whether these models can accurately and ethically capture the experiences of vulnerable populations. In this study, we assess the use of open-source LLMs to inductively code interviews with 21 Black men who have survived community firearm violence. Our results demonstrate that while some configurations of LLMs can identify important codes, overall relevance remains low and is highly sensitive to data processing. Furthermore, LLM guardrails lead to substantial narrative erasure. These findings highlight both the potential and limitations of LLM-assisted qualitative coding and underscore the ethical challenges of applying AI in research involving marginalized communities.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.