TY - GEN
T1 - Fast LiDAR Upsampling using Conditional Diffusion Models
AU - Magnussen Helgesen, Sander Elias
AU - Nakashima, Kazuto
AU - Tørresen, Jim
AU - Kurazume, Ryo
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/8
Y1 - 2024/8
N2 - The search for refining 3D LiDAR data has attracted growing interest motivated by recent techniques such as supervised learning or generative model-based methods. Existing approaches have shown the possibilities for using diffusion models to generate refined LiDAR data with high fidelity, although the performance and speed of such methods have been limited. These limitations make it difficult to execute in real-time, causing the approaches to struggle in real-world tasks such as autonomous navigation and human-robot interaction. In this work, we introduce a novel approach based on conditional diffusion models for fast and high-quality sparse-to-dense upsampling of 3D scene point clouds through an image representation. Our method employs denoising diffusion probabilistic models trained with conditional inpainting masks, which have been shown to give high performance on image completion tasks. We introduce a series of experiments, including multiple datasets, sampling steps, and conditional masks. This paper illustrates that our method outperforms the baselines in sampling speed and quality on upsampling tasks using the KITTI-360 dataset. Furthermore, we illustrate the generalization ability of our approach by simultaneously training on real-world and synthetic datasets, introducing variance in quality and environments.
AB - The search for refining 3D LiDAR data has attracted growing interest motivated by recent techniques such as supervised learning or generative model-based methods. Existing approaches have shown the possibilities for using diffusion models to generate refined LiDAR data with high fidelity, although the performance and speed of such methods have been limited. These limitations make it difficult to execute in real-time, causing the approaches to struggle in real-world tasks such as autonomous navigation and human-robot interaction. In this work, we introduce a novel approach based on conditional diffusion models for fast and high-quality sparse-to-dense upsampling of 3D scene point clouds through an image representation. Our method employs denoising diffusion probabilistic models trained with conditional inpainting masks, which have been shown to give high performance on image completion tasks. We introduce a series of experiments, including multiple datasets, sampling steps, and conditional masks. This paper illustrates that our method outperforms the baselines in sampling speed and quality on upsampling tasks using the KITTI-360 dataset. Furthermore, we illustrate the generalization ability of our approach by simultaneously training on real-world and synthetic datasets, introducing variance in quality and environments.
UR - http://www.scopus.com/inward/record.url?scp=85208035837&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85208035837&partnerID=8YFLogxK
U2 - 10.1109/RO-MAN60168.2024.10731454
DO - 10.1109/RO-MAN60168.2024.10731454
M3 - Conference contribution
AN - SCOPUS:85208035837
T3 - IEEE International Workshop on Robot and Human Communication, RO-MAN
SP - 272
EP - 277
BT - 33rd IEEE International Conference on Robot and Human Interactive Communication, ROMAN 2024
PB - IEEE Computer Society
T2 - 33rd IEEE International Conference on Robot and Human Interactive Communication, ROMAN 2024
Y2 - 26 August 2024 through 30 August 2024
ER -