Do Visual-Language Maps Capture Latent Semantics?

Tutkimustuotos: Muu tuotosScientific

Abstrakti

Visual-language models (VLMs) have recently been introduced in robotic mapping by using the latent representations, i.e., embeddings, of the VLMs to represent the natural language semantics in the map. The main benefit is moving beyond a small set of human-created labels toward open-vocabulary scene understanding. While there is anecdotal evidence that maps built this way support downstream tasks, such as navigation, rigorous analysis of the quality of the maps using these embeddings is lacking. We investigate two critical properties of map quality: queryability and consistency. The evaluation of queryability addresses the ability to retrieve information from the embeddings. We investigate two aspects of consistency: intra-map consistency and inter-map consistency. Intra-map consistency captures the ability of the embeddings to represent abstract semantic classes, and inter-map consistency captures the generalization properties of the representation. In this paper, we propose a way to analyze the quality of maps created using VLMs, which forms an open-source benchmark to be used when proposing new open-vocabulary map representations. We demonstrate the benchmark by evaluating the maps created by two state-of-the-art methods, VLMaps and OpenScene, using two encoders, LSeg and OpenSeg, using real-world data from the Matterport3D data set. We find that OpenScene outperforms VLMaps with both encoders, and LSeg outperforms OpenSeg with both methods.
AlkuperäiskieliEnglanti
DOI - pysyväislinkit
TilaJätetty - 2024
OKM-julkaisutyyppiEi oikeutettu

Sormenjälki

Sukella tutkimusaiheisiin 'Do Visual-Language Maps Capture Latent Semantics?'. Ne muodostavat yhdessä ainutlaatuisen sormenjäljen.

Siteeraa tätä