Big idea: Make image-making AIs stop, think, check, and fix their own work so they get better at both creating pictures and understanding them.
Computer-using agents kept forgetting important visual details over long tasks and could not reliably find up-to-date, step-by-step help for unfamiliar apps.
LLMs can look confident but still change their answers when the surrounding text nudges them, showing that confidence alone isnβt real truthfulness.
This paper shows a new way (called RISE) to find and control how AI models think without needing any human-made labels.