no code implementations • 28 Mar 2024 • Jiacui Huang, Hongtao Zhang, Mingbo Zhao, Zhou Wu
To address this challenge, we propose a new method, namely, Instance-aware Visual Language Map (IVLMap), to empower the robot with instance-level and attribute-level semantic mapping, where it is autonomously constructed by fusing the RGBD video data collected from the robot agent with special-designed natural language map indexing in the bird's-in-eye view.