This placeholder project explores how large-scale models can support manipulation through reusable skills, structured task abstractions, and more transferable control interfaces.
We are especially interested in how vision-language priors and action representations can improve generalization across tasks, objects, and environments.
This page is a placeholder for future project details, papers, demos, and datasets.