The paper asks a simple question: when an AI sees a picture and some text but the instructions say 'only trust the picture,' how does it decide which one to follow?
Large language models (LLMs) donβt act as a single brain; inside, each layer and module quietly makes its own mini-decisions called internal policies.