LLMs for Qualitative Data Analysis Fail on Security-specificComments in Human Experiments

📰 ArXiv cs.AI

arXiv:2604.10834v1 Announce Type: cross Abstract: [Background:] Thematic analysis of free-text justifications in human experiments provides significant qualitative insights. Yet, it is costly because reliable annotations require multiple domain experts. Large language models (LLMs) seem ideal candidates to replace human annotators. [Problem:] Coding security-specific aspects (code identifiers mentioned, lines-of-code mentioned, security keywords mentioned) may require deeper contextual understan

Published 14 Apr 2026
Read full paper → ← Back to Reads