<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en"><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="/feed.xml" rel="self" type="application/atom+xml" /><link href="/" rel="alternate" type="text/html" hreflang="en" /><updated>2026-04-04T03:06:04+00:00</updated><id>/feed.xml</id><title type="html">VU Lab</title><subtitle>The demo site for Bulma Clean Theme, made for Jekyll and GitHub pages websites
</subtitle><entry><title type="html">GSMem: 3D Gaussian Splatting as Persistent Spatial Memory for Zero-Shot Embodied Exploration and Reasoning</title><link href="/highlights/gsmem/" rel="alternate" type="text/html" title="GSMem: 3D Gaussian Splatting as Persistent Spatial Memory for Zero-Shot Embodied Exploration and Reasoning" /><published>2026-04-03T00:00:00+00:00</published><updated>2026-04-03T00:00:00+00:00</updated><id>/highlights/highlight-GSMem</id><content type="html" xml:base="/highlights/gsmem/"><![CDATA[<p>This placeholder project studies how embodied agents can combine vision, language, and action context to build richer scene representations in unstructured environments.</p>

<p>Current directions include long-tail object understanding, semantic grounding under ambiguity, and robust multimodal fusion for agents that must act with incomplete observations.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="highlights" /><category term="Embodied AI" /><category term="3D Vision" /><summary type="html"><![CDATA[This placeholder project studies how embodied agents can combine vision, language, and action context to build richer scene representations in unstructured environments.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-01.svg" /><media:content medium="image" url="/img/research/research-placeholder-01.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Spatial Memory for Long-Horizon Embodied Agents</title><link href="/highlights/2026/04/02/highlight-spatial-memory-agents/" rel="alternate" type="text/html" title="Spatial Memory for Long-Horizon Embodied Agents" /><published>2026-04-02T00:00:00+00:00</published><updated>2026-04-02T00:00:00+00:00</updated><id>/highlights/2026/04/02/highlight-spatial-memory-agents</id><content type="html" xml:base="/highlights/2026/04/02/highlight-spatial-memory-agents/"><![CDATA[<p>This placeholder highlight captures a research direction around persistent world memory for embodied systems.</p>

<p>The goal is to let an agent remember what it has seen, retrieve relevant experiences, and use those memories to guide future actions.</p>

<p>This page can later be replaced with project updates, results, videos, or paper links.</p>]]></content><author><name>VU Lab</name></author><category term="highlights" /><category term="Spatial Memory" /><category term="Embodied AI" /><category term="Long-Horizon Planning" /><summary type="html"><![CDATA[This placeholder highlight captures a research direction around persistent world memory for embodied systems.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/highlights/highlight-02.svg" /><media:content medium="image" url="/img/highlights/highlight-02.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Visual Understanding Benchmark for Open-World Scenes</title><link href="/highlights/2026/04/01/highlight-visual-understanding-benchmark/" rel="alternate" type="text/html" title="Visual Understanding Benchmark for Open-World Scenes" /><published>2026-04-01T00:00:00+00:00</published><updated>2026-04-01T00:00:00+00:00</updated><id>/highlights/2026/04/01/highlight-visual-understanding-benchmark</id><content type="html" xml:base="/highlights/2026/04/01/highlight-visual-understanding-benchmark/"><![CDATA[<p>This placeholder highlight summarizes a lab effort focused on benchmarking visual understanding systems under realistic scene complexity.</p>

<p>The project studies how perception models behave when scenes contain clutter, rare objects, ambiguous language, and shifting context.</p>

<p>We use this page as a placeholder for future highlight content on datasets, evaluation protocols, and model analysis.</p>]]></content><author><name>VU Lab</name></author><category term="highlights" /><category term="Benchmarking" /><category term="Visual Understanding" /><category term="Open-World Perception" /><summary type="html"><![CDATA[This placeholder highlight summarizes a lab effort focused on benchmarking visual understanding systems under realistic scene complexity.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/highlights/highlight-01.svg" /><media:content medium="image" url="/img/highlights/highlight-01.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">GSMem: 3D Gaussian Splatting as Persistent Spatial Memory for Zero-Shot Embodied Exploration and Reasoning</title><link href="/research/multimodal-scene-understanding-embodied-systems/" rel="alternate" type="text/html" title="GSMem: 3D Gaussian Splatting as Persistent Spatial Memory for Zero-Shot Embodied Exploration and Reasoning" /><published>2026-03-30T00:00:00+00:00</published><updated>2026-03-30T00:00:00+00:00</updated><id>/research/GSMem</id><content type="html" xml:base="/research/multimodal-scene-understanding-embodied-systems/"><![CDATA[<p>This placeholder project studies how embodied agents can combine vision, language, and action context to build richer scene representations in unstructured environments.</p>

<p>Current directions include long-tail object understanding, semantic grounding under ambiguity, and robust multimodal fusion for agents that must act with incomplete observations.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="Embodied AI" /><category term="3D Vision" /><summary type="html"><![CDATA[This placeholder project studies how embodied agents can combine vision, language, and action context to build richer scene representations in unstructured environments.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-01.svg" /><media:content medium="image" url="/img/research/research-placeholder-01.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Reconstruction Matters: Learning Geometry-Aligned BEV Representation through 3D Gaussian Splatting</title><link href="/research/Splat2BEV/" rel="alternate" type="text/html" title="Reconstruction Matters: Learning Geometry-Aligned BEV Representation through 3D Gaussian Splatting" /><published>2026-03-29T00:00:00+00:00</published><updated>2026-03-29T00:00:00+00:00</updated><id>/research/Splat2BEV</id><content type="html" xml:base="/research/Splat2BEV/"><![CDATA[<p>This placeholder project focuses on how embodied agents maintain useful spatial memory over long time horizons while reasoning about goals, constraints, and uncertainty.</p>

<p>We are interested in navigation policies, memory-augmented world models, and planning systems that remain effective when tasks require multi-step reasoning across large spaces.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="3D Vision" /><category term="Autonomous Driving" /><summary type="html"><![CDATA[This placeholder project focuses on how embodied agents maintain useful spatial memory over long time horizons while reasoning about goals, constraints, and uncertainty.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-02.svg" /><media:content medium="image" url="/img/research/research-placeholder-02.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Robust 3D Mapping and Adaptive Navigation</title><link href="/research/robust-3d-mapping-adaptive-navigation/" rel="alternate" type="text/html" title="Robust 3D Mapping and Adaptive Navigation" /><published>2026-03-28T00:00:00+00:00</published><updated>2026-03-28T00:00:00+00:00</updated><id>/research/robust-3d-mapping-adaptive-navigation</id><content type="html" xml:base="/research/robust-3d-mapping-adaptive-navigation/"><![CDATA[<p>This placeholder project explores how agents can build stable 3D maps while environments evolve, sensing degrades, or scene geometry changes over time.</p>

<p>Representative directions include geometry-aware learning, map refinement under uncertainty, and adaptive navigation policies that remain effective in dynamic real-world deployments.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="SLAM" /><category term="Navigation" /><category term="Project Overview" /><summary type="html"><![CDATA[This placeholder project explores how agents can build stable 3D maps while environments evolve, sensing degrades, or scene geometry changes over time.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-03.svg" /><media:content medium="image" url="/img/research/research-placeholder-03.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Video-Language Grounding for Open-World Agents</title><link href="/research/video-language-grounding-open-world-agents/" rel="alternate" type="text/html" title="Video-Language Grounding for Open-World Agents" /><published>2026-03-27T00:00:00+00:00</published><updated>2026-03-27T00:00:00+00:00</updated><id>/research/video-language-grounding-open-world-agents</id><content type="html" xml:base="/research/video-language-grounding-open-world-agents/"><![CDATA[<p>This placeholder project examines how agents align visual observations with language over time, especially when scenes, objects, and goals evolve beyond closed-set assumptions.</p>

<p>We are interested in open-world recognition, grounded language understanding, and long-horizon video reasoning for systems that operate continuously rather than on isolated clips.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="Vision-Language" /><category term="Recognition" /><category term="Project Overview" /><summary type="html"><![CDATA[This placeholder project examines how agents align visual observations with language over time, especially when scenes, objects, and goals evolve beyond closed-set assumptions.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-04.svg" /><media:content medium="image" url="/img/research/research-placeholder-04.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Foundation Models for Robotic Manipulation</title><link href="/research/foundation-models-robotic-manipulation/" rel="alternate" type="text/html" title="Foundation Models for Robotic Manipulation" /><published>2026-03-26T00:00:00+00:00</published><updated>2026-03-26T00:00:00+00:00</updated><id>/research/foundation-models-robotic-manipulation</id><content type="html" xml:base="/research/foundation-models-robotic-manipulation/"><![CDATA[<p>This placeholder project explores how large-scale models can support manipulation through reusable skills, structured task abstractions, and more transferable control interfaces.</p>

<p>We are especially interested in how vision-language priors and action representations can improve generalization across tasks, objects, and environments.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="Foundation Models" /><category term="Manipulation" /><category term="Project Overview" /><summary type="html"><![CDATA[This placeholder project explores how large-scale models can support manipulation through reusable skills, structured task abstractions, and more transferable control interfaces.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-05.svg" /><media:content medium="image" url="/img/research/research-placeholder-05.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Uncertainty Estimation and Robust Deployment</title><link href="/research/uncertainty-estimation-robust-deployment/" rel="alternate" type="text/html" title="Uncertainty Estimation and Robust Deployment" /><published>2026-03-25T00:00:00+00:00</published><updated>2026-03-25T00:00:00+00:00</updated><id>/research/uncertainty-estimation-robust-deployment</id><content type="html" xml:base="/research/uncertainty-estimation-robust-deployment/"><![CDATA[<p>This placeholder project investigates how autonomous systems should estimate confidence, react to uncertainty, and remain reliable when real-world conditions depart from training assumptions.</p>

<p>Representative themes include distribution shift detection, calibrated decision making, and risk-aware inference pipelines for perception and control.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="Uncertainty" /><category term="Robustness" /><category term="Project Overview" /><summary type="html"><![CDATA[This placeholder project investigates how autonomous systems should estimate confidence, react to uncertainty, and remain reliable when real-world conditions depart from training assumptions.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-06.svg" /><media:content medium="image" url="/img/research/research-placeholder-06.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Interactive 3D World Models for Embodied Training</title><link href="/research/interactive-3d-world-models-embodied-training/" rel="alternate" type="text/html" title="Interactive 3D World Models for Embodied Training" /><published>2026-03-24T00:00:00+00:00</published><updated>2026-03-24T00:00:00+00:00</updated><id>/research/interactive-3d-world-models</id><content type="html" xml:base="/research/interactive-3d-world-models-embodied-training/"><![CDATA[<p>This placeholder project focuses on compact world representations that can support simulation, future prediction, and scalable training of embodied agents.</p>

<p>We are interested in interactive 3D environments, controllable simulators, and learned world models that bridge synthetic training and real deployment.</p>

<p>This page is a placeholder for future project details, papers, demos, and datasets.</p>]]></content><author><name>VU Lab</name></author><category term="research" /><category term="World Models" /><category term="Simulation" /><category term="Project Overview" /><summary type="html"><![CDATA[This placeholder project focuses on compact world representations that can support simulation, future prediction, and scalable training of embodied agents.]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="/img/research/research-placeholder-07.svg" /><media:content medium="image" url="/img/research/research-placeholder-07.svg" xmlns:media="http://search.yahoo.com/mrss/" /></entry></feed>