Back

Probing Hidden States for Calibrated, Alignment-Resistant Predictions in LLMs

2025-09-19 health informatics Title + abstract only
View on medRxiv
Show abstract

Scientific applications of large language models (LLMs) demand reliable, well-calibrated predictions, but standard generative approaches often fail to fully access relevant knowledge contained in their internal representations. As a result, models appear less capable than they are, with useful information remaining latent. We present PING (Probing INternal states of Generative models), an open-source framework that trains lightweight probes on frozen, HuggingFace-compatible transformers to deliv...

Predicted journal destinations