Inside the Forward Pass: Can Transformer Internals Predict Correctness?
TL;DR: Internal transformer signals (entropy, attention, hidden state statistics) predict generation correctness with AUROC 0.60–0.90 under grouped held-out evaluation, without looking at the output text. The first 10 generated tokens carry most of the predictive signal for code tasks. Model confidence scores are nearly uncorrelated with correctness for…