GSMem: 3D Gaussian Splatting as Persistent Spatial Memory for Zero-Shot Embodied Exploration and Reasoning

This placeholder project studies how embodied agents can combine vision, language, and action context to build richer scene representations in unstructured environments.

Current directions include long-tail object understanding, semantic grounding under ambiguity, and robust multimodal fusion for agents that must act with incomplete observations.

This page is a placeholder for future project details, papers, demos, and datasets.

Embodied AI 3D Vision