Utonia is a single brain (encoder) that learns from many kinds of 3D point clouds, like indoor rooms, outdoor streets, tiny toys, and even city maps.
NE-Dreamer is a model-based reinforcement learning agent that skips rebuilding pixels and instead learns by predicting the next step’s hidden features.
The paper shows that when we train with the popular InfoNCE contrastive loss, the learned features start to behave like they come from a Gaussian (bell-shaped) distribution.
BatCoder teaches a code model to write both code and its documentation by doing a round trip: from code to docs and back to code.
This paper builds a fair, big playground (a benchmark) to test many EEG foundation models side-by-side on the same rules.
Co2S is a new way to train segmentation models with very few labels by letting two different students (CLIP and DINOv3) learn together and correct each other.
Pixels are the raw stuff of images, and this paper shows you can learn great vision skills by predicting pixels directly, not by comparing fancy hidden features.
The paper tackles a paradox: visual tokenizers that get great pixel reconstructions often make worse images when used for generation.