This study presents a robust approach to reconstructing a three dimensional (3-D) translucent object using a single time-of-flight depth camera with simple user marks. Because the appearance of translucent objects depends on the light interaction with the surrounding environment, the measurement using depth cameras is considerably biased or invalid. Although several existing methods attempt to model the depth error of translucent objects, their model remains partial because of object assumptions and its sensitivity to noise. In this study, we introduce a ground plane and piece-wise linear surface model as priors and construct a robust 3-D reconstruction framework for translucent objects. These two depth priors are combined with the depth error model built on the time-of-flight principle. Extensive evaluation of various real data reveals that the proposed method substantially improves the accuracy and reliability of 3-D reconstruction for translucent objects.
Bibliographical noteFunding Information:
The MSIP (Ministry of Science, ICT and Future Planning), Korea, under the "ICT Consilience Creative Program" (IITP-R0346-16-1008) supervised by the IITP (Institute for Information and communications Technology Promotion); the Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Science, ICT and Future Planning (NRF-2016R1A2B4016236).
© 2017 Optical Society of America.
All Science Journal Classification (ASJC) codes
- Atomic and Molecular Physics, and Optics