How to Detect Test Set Contamination in Black‑Box Language Models
Researchers propose a black‑box method to expose test‑set leakage in large language models by comparing log‑probability shifts when test items are shuffled, using Monte‑Carlo estimation and a sharded likelihood test, and demonstrate its effectiveness on several models including Mistral‑7B.
