TY - GEN
T1 - Learning Neural Volumetric Field for Point Cloud Geometry Compression
AU - Hu, Yueyu
AU - Wang, Yao
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Due to the diverse sparsity, high dimensionality, and large temporal variation of dynamic point clouds, it remains a challenge to design an efficient point cloud compression method. We propose to code the geometry of a given point cloud by learning a neural volumetric field. Instead of representing the entire point cloud using a single overfit network, we divide the entire space into small cubes and represent each non-empty cube by a neural network and an input latent code. The network is shared among all the cubes in a single frame or multiple frames, to exploit the spatial and temporal redundancy. The neural field representation of the point cloud includes the network parameters and all the latent codes, which are generated by using back-propagation over the network parameters and its input. By considering the entropy of the network parameters and the latent codes as well as the distortion between the original and reconstructed cubes in the loss function, we derive a rate-distortion (R-D) optimal representation. Experimental results show that the proposed coding scheme achieves superior R-D performances compared to the octree-based G-PCC, especially when applied to multiple frames of a point cloud video. The code is available at https://github.com/huzi96/NVFPCC/.
AB - Due to the diverse sparsity, high dimensionality, and large temporal variation of dynamic point clouds, it remains a challenge to design an efficient point cloud compression method. We propose to code the geometry of a given point cloud by learning a neural volumetric field. Instead of representing the entire point cloud using a single overfit network, we divide the entire space into small cubes and represent each non-empty cube by a neural network and an input latent code. The network is shared among all the cubes in a single frame or multiple frames, to exploit the spatial and temporal redundancy. The neural field representation of the point cloud includes the network parameters and all the latent codes, which are generated by using back-propagation over the network parameters and its input. By considering the entropy of the network parameters and the latent codes as well as the distortion between the original and reconstructed cubes in the loss function, we derive a rate-distortion (R-D) optimal representation. Experimental results show that the proposed coding scheme achieves superior R-D performances compared to the octree-based G-PCC, especially when applied to multiple frames of a point cloud video. The code is available at https://github.com/huzi96/NVFPCC/.
KW - Neural Field
KW - Point Cloud Compression
KW - Rate-Distortion optimization
UR - http://www.scopus.com/inward/record.url?scp=85147664486&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147664486&partnerID=8YFLogxK
U2 - 10.1109/PCS56426.2022.10018029
DO - 10.1109/PCS56426.2022.10018029
M3 - Conference contribution
AN - SCOPUS:85147664486
T3 - 2022 Picture Coding Symposium, PCS 2022 - Proceedings
SP - 127
EP - 131
BT - 2022 Picture Coding Symposium, PCS 2022 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 Picture Coding Symposium, PCS 2022
Y2 - 7 December 2022 through 9 December 2022
ER -