BEACON: Language-Conditioned Navigation Affordance Prediction under Occlusion

PDF Viewer

BUILDER'S SANDBOX

Build This Paper

Use an AI coding agent to implement this research.

OpenAI Codex
OpenAI CodexAI Agent

Lightweight coding agent in your terminal.

Claude Code
Claude CodeAI Agent

Agentic coding tool for terminal workflows.

AntiGravity IDE
AntiGravity IDEScaffolding

AI agent mindset installer and workflow scaffolder.

Cursor
CursorIDE

AI-first code editor built on VS Code.

VS Code
VS CodeIDE

Free, open-source editor by Microsoft.

Estimated $9K - $13K over 6-10 weeks.

See exactly what it costs to build this -- with 3 comparable funded startups.

7-day free trial. Cancel anytime.

Discover the researchers behind this paper and find similar experts.

7-day free trial. Cancel anytime.

References (36)

[1]
DrivePI: Spatial-aware 4D MLLM for Unified Autonomous Driving Understanding, Perception, Prediction and Planning
2025Zhe Liu, Runhu Huang et al.
[2]
RoboAfford++: A Generative AI-Enhanced Dataset for Multimodal Affordance Learning in Robotic Manipulation and Navigation
2025Xiaoshuai Hao, Yingbo Tang et al.
[3]
RoboAfford: A Dataset and Benchmark for Enhancing Object and Spatial Affordance Learning in Robot Manipulation
2025Yingbo Tang, Lingfeng Zhang et al.
[4]
More than A Point: Capturing Uncertainty with Adaptive Affordance Heatmaps for Spatial Grounding in Robotic Tasks
2025Xinyu Shao, Yanzhe Tang et al.
[5]
3D Aware Region Prompted Vision Language Model
2025An-Chieh Cheng, Yang Fu et al.
[6]
Spatial Reasoning with Vision-Language Models in Ego-Centric Multi-View Scenes
2025Mohsen Gholami, Ahmad Rezaei et al.
[7]
RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics
2025Enshen Zhou, Jingkun An et al.
[8]
TrackVLA: Embodied Visual Tracking in the Wild
2025Shaoan Wang, Jiazhao Zhang et al.
[9]
BEVDriver: Leveraging BEV Maps in LLMs for Robust Closed-Loop Driving
2025Katharina Winter, Mark Azer et al.
[10]
SpatialVLA: Exploring Spatial Representations for Visual-Language-Action Model
2025Delin Qu, Haoming Song et al.
[11]
SpatialCoT: Advancing Spatial Reasoning through Coordinate Alignment and Chain-of-Thought for Embodied Task Planning
2025Yuecheng Liu, Dafeng Chi et al.
[12]
BEVFormer: Learning Bird’s-Eye-View Representation From LiDAR-Camera via Spatiotemporal Transformers
2024Zhiqi Li, Wenhai Wang et al.
[13]
RoboSpatial: Teaching Spatial Understanding to 2D and 3D Vision-Language Models for Robotics
2024Chan Hee Song, Valts Blukis et al.
[14]
From Cognition to Precognition: A Future-Aware Framework for Social Navigation
2024Zeying Gong, Tianshuai Hu et al.
[15]
SpatialBot: Precise Spatial Understanding with Vision Language Models
2024Wenxiao Cai, Yaroslav Ponomarenko et al.
[16]
RoboPoint: A Vision-Language Model for Spatial Affordance Prediction for Robotics
2024Wentao Yuan, Jiafei Duan et al.
[17]
SpatialRGPT: Grounded Spatial Reasoning in Vision Language Model
2024An-Chieh Cheng, Hongxu Yin et al.
[18]
Vision-based 3D occupancy prediction in autonomous driving: a review and outlook
2024Yanan Zhang, Jinqing Zhang et al.
[19]
SceneSense: Diffusion Models for 3D Occupancy Synthesis from Partial Observation
2024Alec Reed, Brendan Crowe et al.
[20]
LINGO-Space: Language-Conditioned Incremental Grounding for Space
2024Dohyun Kim, Nayoung Oh et al.

Showing 20 of 36 references

Founder's Pitch

"BEACON enhances robot navigation by predicting traversable locations in occluded environments using language instructions and depth data."

Robotics NavigationScore: 8View PDF ↗

Commercial Viability Breakdown

0-10 scale

High Potential

3/4 signals

7.5

Quick Build

2/4 signals

5

Series A Potential

3/4 signals

7.5

Sources used for this analysis

arXiv Paper

Full-text PDF analysis of the research paper

GitHub Repository

Code availability, stars, and contributor activity

Citation Network

Semantic Scholar citations and co-citation patterns

Community Predictions

Crowd-sourced unicorn probability assessments

Analysis model: GPT-4o · Last scored: 3/10/2026

Explore the full citation network and related research.

7-day free trial. Cancel anytime.

Understand the commercial significance and market impact.

7-day free trial. Cancel anytime.

Get detailed profiles of the research team.

7-day free trial. Cancel anytime.

Related Papers

Loading…