A Metamorphic Testing Approach to Diagnosing Memorization in LLM-Based Program Repair

📰 ArXiv cs.AI

arXiv:2604.21579v1 Announce Type: cross Abstract: LLM-based automated program repair (APR) techniques have shown promising results in reducing debugging costs. However, prior results can be affected by data leakage: large language models (LLMs) may memorize bug fixes when evaluation benchmarks overlap with their pretraining data, leading to inflated performance estimates. In this paper, we investigate whether we can better reveal data leakage by combining metamorphic testing (MT) with negative l

Published 25 Apr 2026
Read full paper → ← Back to Reads