MMR-Life is a new test (benchmark) that checks how AI understands everyday situations using several real photos at once.
The paper fixes a common problem in AI: models can read pictures and text well, but they often mess up the logic behind them.
Modern image editors can now follow visual prompts like arrows and scribbles, which opens a new way for attackers to hide harmful instructions inside images.
The paper studies how to make and judge scientific images that are not just pretty but scientifically correct.
ATLAS is a system that picks the best mix of AI models and helper tools for each question, instead of using just one model or a fixed tool plan.
Real people often ask vague questions with pictures, and today’s vision-language models (VLMs) struggle with them.