LiLMaps: Learnable Implicit Language Maps

One of the current trends in robotics is to employ large language models (LLMs) to provide non-predefined command execution and natural human-robot interaction. It is useful to have an environment map together with its language representation, which can be further utilized by LLMs. Such a comprehensive scene representation enables numerous ways of interaction with the map for autonomously operating robots. In this work, we present an approach that enhances incremental implicit mapping through the integration of vision-language features. Specifically , we (i) propose a decoder optimization technique for implicit language maps which can be used when new objects appear on the scene, and (ii) address the problem of inconsistent vision-language predictions between different viewing positions. Our experiments demonstrate the effectiveness of LiLMaps and solid improvements in performance .

Informationen zur Zitierung

Kruzhkov, Evgenii; Behnke, Sven: LiLMaps: Learnable Implicit Language Maps, IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2025, https://www.researchgate.net/publication/387488336_LiLMaps_Learnable_Implicit_Language_Maps, Kruzhkov.Behnke.2025a,

Assoziierte Lamarr-ForscherInnen

lamarr institute person Behnke Sven - Lamarr Institute for Machine Learning (ML) and Artificial Intelligence (AI)

Prof. Dr. Sven Behnke

Area Chair Embodied AI zum Profil