Abstract
Input-level domain adaptation reduces the burden of a neural encoder without supervision by reducing the domain gap at the input level. Input-level domain adaptation is widely employed in 2D visual domain, e.g., images and videos, but is not utilized for 3D point clouds. We propose the use of input-level domain adaptation for 3D point clouds, namely, point-level domain adaptation. Specifically, we propose to learn a transformation of 3D point clouds by searching the best combination of operations on point clouds that transfer data from the source domain to the target domain while maintaining the classification label without supervision of the target label. We decompose the learning objective into two terms, resembling domain shift and preserving label information. On the PointDA-10 benchmark dataset, our method outperforms state-of-the-art, unsupervised, point cloud domain adaptation methods by large margins (up to + 3.97 % in average).
Original language | English |
---|---|
Pages (from-to) | 56901-56913 |
Number of pages | 13 |
Journal | IEEE Access |
Volume | 10 |
DOIs | |
Publication status | Published - 2022 |
Bibliographical note
Funding Information:This work was supported by the Institute for Information & Communications Technology Promotion (IITP) Grant funded by the Korean Government [Ministry of Science and ICT (MSIT)] (Development of Ultra Low-Power Mobile Deep Learning Semiconductor With Compression/Decompression of Activation/Kernel Data) under Grant 2019-0-01351.
Publisher Copyright:
© 2013 IEEE.
All Science Journal Classification (ASJC) codes
- Computer Science(all)
- Materials Science(all)
- Engineering(all)
- Electrical and Electronic Engineering