Kimi K2.5 is a new open-source AI that can read both text and visuals (images and videos) and act like a team of helpers to finish big tasks faster.
XR is a new, training-free team of AI helpers that finds images using both a reference picture and a short text edit (like “same jacket but red”).
Real people often ask vague questions with pictures, and today’s vision-language models (VLMs) struggle with them.
ThinkRL-Edit teaches an image editor to think first and draw second, which makes tricky, reasoning-heavy edits much more accurate.
The paper builds YearGuessr, a giant, worldwide photo-and-text dataset of 55,546 buildings with their construction years (1001–2024), GPS, and popularity (page views).