TY - GEN
T1 - MuSHRoom : Multi-Sensor Hybrid Room Dataset for Joint 3D Reconstruction and Novel View Synthesis
AU - Ren, Xuqian
AU - Wang, Wenjia
AU - Cai, Dingding
AU - Tuominen, Tuuli
AU - Kannala, Juho
AU - Rahtu, Esa
N1 - Publisher Copyright: © 2024 IEEE.
PY - 2024/1/3
Y1 - 2024/1/3
N2 - Metaverse technologies demand accurate, real-time, and immersive modeling on consumer-grade hardware for both non-human perception (e.g., drone/robot/autonomous car navigation) and immersive technologies like AR/VR, requiring both structural accuracy and photorealism. However, there exists a knowledge gap in how to apply geometric reconstruction and photorealism modeling (novel view synthesis) in a unified framework. To address this gap and promote the development of robust and immersive modeling and rendering with consumer-grade devices, first, we propose a real-world Multi-Sensor Hybrid Room Dataset (MuSHRoom). Our dataset presents exciting challenges and requires state-of-the-art methods to be cost-effective, robust to noisy data and devices, and can jointly learn 3D reconstruction and novel view synthesis instead of treating them as separate tasks, making them ideal for realworld applications. Second, we benchmark several famous pipelines on our dataset for joint 3D mesh reconstruction and novel view synthesis. Finally, in order to further improve the overall performance, we propose a new method that achieves a good trade-off between the two tasks. Our dataset and benchmark show great potential in promoting the improvements for fusing 3D reconstruction and highquality rendering in a robust and computationally efficient end-to-end fashion. The dataset and code are available at the project website: https://xuqianren.github.io/publications/MuSHRoom/.
AB - Metaverse technologies demand accurate, real-time, and immersive modeling on consumer-grade hardware for both non-human perception (e.g., drone/robot/autonomous car navigation) and immersive technologies like AR/VR, requiring both structural accuracy and photorealism. However, there exists a knowledge gap in how to apply geometric reconstruction and photorealism modeling (novel view synthesis) in a unified framework. To address this gap and promote the development of robust and immersive modeling and rendering with consumer-grade devices, first, we propose a real-world Multi-Sensor Hybrid Room Dataset (MuSHRoom). Our dataset presents exciting challenges and requires state-of-the-art methods to be cost-effective, robust to noisy data and devices, and can jointly learn 3D reconstruction and novel view synthesis instead of treating them as separate tasks, making them ideal for realworld applications. Second, we benchmark several famous pipelines on our dataset for joint 3D mesh reconstruction and novel view synthesis. Finally, in order to further improve the overall performance, we propose a new method that achieves a good trade-off between the two tasks. Our dataset and benchmark show great potential in promoting the improvements for fusing 3D reconstruction and highquality rendering in a robust and computationally efficient end-to-end fashion. The dataset and code are available at the project website: https://xuqianren.github.io/publications/MuSHRoom/.
KW - 3D computer vision
KW - Algorithms
KW - Computational photography
KW - Datasets and evaluations
KW - image and video synthesis
UR - http://www.scopus.com/inward/record.url?scp=85191982650&partnerID=8YFLogxK
U2 - 10.1109/WACV57701.2024.00445
DO - 10.1109/WACV57701.2024.00445
M3 - Conference article in proceedings
AN - SCOPUS:85191982650
T3 - IEEE Winter Conference on Applications of Computer Vision
SP - 4496
EP - 4505
BT - Proceedings - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024
PB - IEEE
T2 - IEEE Winter Conference on Applications of Computer Vision
Y2 - 4 January 2024 through 8 January 2024
ER -