Sergey Levine was just presenting in the Exploration in AI @ #ICML2025 and promoted that exploration needs to be grounded, and that VLMs are a good source ;-) Check our paper below
๐
Sergey Levine was just presenting in the Exploration in AI @ #ICML2025 and promoted that exploration needs to be grounded, and that VLMs are a good source ;-) Check our paper below
๐
Want to find out more about SENSEI?
๐ฃ๏ธICML Poster West Exhibition Hall, 16 Jul, 11a.m. PDT, No. W-707
๐arxiv.org/abs/2503.01584
๐sites.google.com/view/sensei-paper
Work done with @cgumbsch.bsky.social (co-first), @zadaianchuk.bsky.social, @pavelkolevbg.bsky.social and @gmartius.bsky.social
8/8
SENSEI can also guide exploration in combination with task rewards. When playing Pokรฉmon Red from pixels, we achieve superior performance to Dreamer (pure task rewards) and Plan2Explore. Only SENSEI manages to obtain the first Gym Badge within 2M steps of exploration ๐ฅ
7/8
The agent learns a world model during exploration that can later be re-used to solve downstream tasks. We demonstrate more sample-efficient policy learning with SENSEI compared to exploration via Plan2Explore.
6/8
Through the combination of semantic exploration with epistemic uncertainty, the agent unlocks a variety of interesting behaviors during task-free exploration. For example, in Robodesk the agent focuses on interacting with all available objects ๐ฆพ
5/8
To continuously push the frontier of experience, we combine semantic rewards with epistemic uncertainty deploying an adaptive go-explore strategy. The agent first tries to reach interesting situations (๐ semantic reward) and then tries new things from there (๐ uncertainty)
4/8
How do we get a signal for meaningful behavior?๐ค
Our approach is to use human priors found in foundation models. We extend MOTIF to VLMs: A VLM compares observation pairs, collected through self-supervised exploration. This ranking is distilled into a reward function.
3/8
Intrinsically motivated exploration faces a chicken-or-egg problem: how do you know whatโs worth exploring before trying it out and experiencing the consequences?
Children solve this by observing and imitating adults. We bring such semantic exploration to artificial agents.
2/8
โจIntroducing SENSEIโจ We bring semantically meaningful exploration to model-based RL using VLMs.
With intrinsic rewards for novel yet useful behaviors, SENSEI showcases strong exploration in MiniHack, Pokรฉmon Red & Robodesk.
Accepted at ICML 2025๐
Joint work with @cgumbsch.bsky.social
๐งต