Papers
1–3 of 3Research Paper·Mar 16, 2026
RealVLG-R1: A Large-Scale Real-World Visual-Language Grounding Benchmark for Robotic Perception and Manipulation
Visual-language grounding aims to establish semantic correspondences between natural language and visual entities, enabling models to accurately identify and localize target objects based on textual i...
9.0 viability
Research Paper·Mar 16, 2026
Coupled Particle Filters for Robust Affordance Estimation
Robotic affordance estimation is challenging due to visual, geometric, and semantic ambiguities in sensory input. We propose a method that disambiguates these signals using two coupled recursive estim...
7.0 viability
Research Paper·Jan 21, 2026
Vision-Language Models on the Edge for Real-Time Robotic Perception
Vision-Language Models (VLMs) enable multimodal reasoning for robotic perception and interaction, but their deployment in real-world systems remains constrained by latency, limited onboard resources, ...
6.0 viability