Multi-resolution intra-predictive coding of 3D point cloud attributes
Abstract
We propose an intra frame predictive strategy for compression of 3D point cloud attributes. Our approach is integrated with the region adaptive graph Fourier transform (RAGFT), a multi-resolution transform formed by a composition of localized block transforms, which produces a set of low pass (approximation) and high pass (detail) coefficients at multiple resolutions. Since the transform operations are spatially localized, RAGFT coefficients at a given resolution may still be correlated. To exploit this phenomenon, we propose an intra-prediction strategy, in which decoded approximation coefficients are used to predict uncoded detail coefficients. The prediction residuals are then quantized and entropy coded. For the 8i dataset, we obtain gains up to db as compared to intra predicted point cloud compresion based on the region adaptive Haar transform (RAHT).
Index Terms— 3D point clouds, intra prediction, multiresolution transform, graph fourier transform
1 Introduction
3D point clouds (3DPC) have become the preferred representation of 3D scenes, people and objects [1, 2, 3]. They consist of a list of points coordinates in 3D space along with color attributes. Recent advancements in real time 3D capture, along with potential applications to entertainment and immersive communications, have prompted research on 3DPC compression [1, 2].
This paper focuses on 3DPC attribute compression. Earlier approaches for compression of 3DPC attributes were based on transform coding techniques, that is, transformation followed by quantization and entropy coding, similar to modern image codecs. For 3DPCs, a popular approach is based on the region adaptive hierarchical (or Haar) transform (RAHT) [4, 5]. In image and video coding, intra and inter prediction are often combined with transform coding. Predictive methods for 3DPC compression have only recently become popular, which may be explained by the fact that good spatial and temporal predictors are harder to obtain for 3DPCs, since these i) represent complex surface and non-surface geometries, ii) have spatial frame to frame irregularity, and iii) lack temporal consistency. There has been substantial recent work to apply transform coding to inter frame (temporal) prediction residuals [3, 6, 7, 8, 9]. However, intra frame prediction is less explored. In [10, 11], block based intra-prediction similar to video was considered. More recently, MPEG adopted an intra prediction strategy for the RAHT (I-RAHT) [12], which uses a multi-resolution prediction, instead of traditional directional block based prediction.


In I-RAHT, low pass decoded RAHT coefficients are used to predict high pass RAHT coefficients. To achieve this, a higher resolution predictor point cloud is constructed from lower resolution decoded coefficients. Each predictor attribute in the high resolution point cloud is obtained as a linear combination of the nearest attributes in the lower resolution point cloud. Then the detail coefficients of this predictor signal are used to predict the target detail coefficients. While this approach to construct a higher resolution point cloud is suitable for the RAHT, since it is formed by a composition of orthogonal transforms, it fails to take into account the more complex geometry of the higher resolution point cloud. This becomes even more important when considering larger block transforms with more points used by the region adaptive graph Fourier transform (RAGFT) [13], as shown in Figure 1.
In this paper we propose an intra-predictive coding framework for RAGFT (I-RAGFT). We use the same multi-resolution prediction strategy used in I-RAHT, where decoded approximation coefficients are used to predict and code detail coefficients, as depicted in Figure 2. However, different from the prediction algorithm used by I-RAHT, our proposed predictors exploits the higher resolution point cloud geometry. We start by projecting the low resolution signal onto the higher resolution geometry by zero padding and applying a one level inverse RAGFT. The resulting interpolated signal is piece-wise constant, as depicted by the colored voxels in Figure 1(b). Then a smoothing graph filter [14, 15, 16, 17] is applied using a graph constructed on top of the higher resolution point cloud. While our approach is new for 3DPCs, similar ideas have been used to improve image coding by predicting wavelet coefficients with learning based super resolution algorithms [18].
The difference between our proposed approach and I-RAHT is illustrated in Figure 1. I-RAHT uses a block-level predictor, since the approximation coefficients of the nearest neighboring blocks are used (Figure 1(a)), while our approach uses a point-level predictor, where fine resolution point values are interpolated, then filtered (across block boundaries), so that only nearby points (instead of nearby blocks) are used to compute the predictor. We show through compression experiments that the proposed approach can outperform I-RAHT, when using uniform quantization and adaptive RLGR entropy coding [19].
2 Multi resolution predictive coding
Consider a point cloud with point coordinates stored in the matrix , and attributes . We will assume the attributes are processed with a multi-resolution transform, which at each resolution level takes an attribute vector , and produces approximation and detail coefficients at resolution
(1) |
The transform matrix is an orthonormal matrix, thus . We assume that the original 3DPC attributes are stored at the highest resolution, , so that . After applying (1) times, we obtain transform coefficients
(2) |
Several orthonormal transforms for 3DPC attributes can be described this way, including the block based graph Fourier transform [20], RAHT [4] and RAGFT [13]. Since RAGFT is a composition of spatially localized block transforms, there may be additional redundancy between transformed coefficients, similar to what is observed for the RAHT [12]. While previous approaches code the coefficients in (2) independently ignoring their dependencies across blocks, in this work we exploit them to improve coding efficiency.
Denote by , the quantized low pass coefficients, where is a quantization operator, and denote by the corresponding decoded coefficients. Now we define several quantities recursively. The decoded approximation coefficient at resolution is given by , while the corresponding detail coefficient at the same resolution is . We will assume there is function , that predicts detail coefficients from approximation coefficients. This predictor will be explained in detail in the next section. Using we compute a residual and quantize it obtaining
(3) |
If the predictor is good enough, coding is more efficient than coding the transform coefficients directly. The decoded details coefficients are given by
(4) |
The decoded approximation and detail coefficients at resolution are used to obtain approximation coefficients at resolution with
(5) |
The quantities can be computed recursively as depicted in Figure 2, starting from the lowest resolution coefficients . A typical transform coding strategy would quantize and entropy encode (2). In this work we encode
(6) |

3 Multi resolution prediction
In this section we describe the function , used to predict the high pass coefficients .
3.1 Graph representation of point clouds
While forming transform coefficients (2), many transforms [4, 13], either explicitely or implicitely, produce sets of point coordinates at various resolutions (e.g., by down-sampling), thus for each resolution , we have a point cloud , where and .
For each of these point clouds consider a graph . The matrix is the adjacency matrix and is the degree matrix. The graph has edge set , where if point is “near” to point . Edge weights are given by
(7) |
3.2 Constructing the predictor
The first step in constructing our predictor is interpolation of the low resolution point cloud by zero padding, and inverse transformation, leading to the attribute signal
(8) |
and higher resolution point cloud . We propose applying a smoothing (low pass) filter to the point cloud , and then applying the transform . The predictor block from Figure 2 is depicted in Figure 3, with the caveat that is ignored.




3.3 Graph filtering of RAGFT coefficients
In the RAGFT [13], each transform coefficient has an importance weight that depends on the size (number of points) of the region of the point cloud they represent. The th point of the point cloud of resolution is denoted by . The point has a set of children at resolution denoted by . The index set of the children is denoted by . We denote the weight of point by . Based on this parent-child relation, the weights can be computed recursively as
(9) |
and the weights at full resolution are for all . For the RAGFT, the interpolation equation (8) has a closed form, thus the th entry of is equal to
(10) |
where point is the parent of . Note that if we define the diagonal matrix of weights by , with entry equal to , then the interpolated signal is
(11) |
where is the indicator of the set . Thus after normalization by the square root of the point weights, the interpolated signal is piece-wise constant. In fact, this signal is constant within cube shaped regions, because the RAGFT is is a composition of localized “block” transforms. Figure 1(b) depicts this piece-wise constant signal. Given that has this form, we take into account the normalization matrix , and the piece-wise constant structure, when designing our smoothing operator. The first step is to normalize the entries of by their point weights using . The resulting piece-wise constant signal is filtered to smooth out boundaries between regions. Finally, we scale back each attribute using the matrix . The proposed filtering operation has the form
(12) |
At resolution , the adjacency matrix is constructed using nearest neighbors on the point cloud geometry . The signal is transformed using (1), resulting in and .
3.4 Complexity
The proposed intra predicted RAGFT can be computed by performing a forward and inverse RAGFT, along with a graph construction and a graph filtering operation per resolution level. For a point cloud with points the forward and inverse RAGFT have complexity . If the octree [21, 22] has to be computed, overall complexity increases to . If the number of points at resolution is equal to , computation of using (11) takes time. Graph construction with nearest neighbors requires operations, while can be computed using sparse matrix vector products in time. If is , then prediction operations at all resolutions have complexity , resulting in an overall complexity of the intra predicted RAGFT.
4 Numerical results
We integrate intra prediction with the RAGFT with blocks at all resolution levels, thus each localized block transform processes at most points. The proposed predictors are implemented using graph filters, that use nearest neighbor graphs, with neighbors per point. We denote this approach by “I-RAGFT”. We also implement predictors similar to those used by the G-PCC implementation of I-RAHT using KNN from the lower resolution graph (see Fig. 1(a)). For this approach we set , found after optimizing for best performance. This approach is denoted by “I-RAGFT LowRes”. We also compare against the RAGFT with high resolution block size equal to , which achieves highest coding performance compared to RAHT, and the RAGFT with high resolution block size equal to , which provides a mild improvement over RAHT (see [13] for details). For RAGFT and intra RAGFT, we uniformly quantize coefficients and entropy code them using adaptive run-length Golomb-Rice (RLGR) algorithm [19].
The state of the art in non-video-based compression of point cloud attributes, called G-PCC [5] uses I-RAHT. Several techniques are implemented in G-PCC to improve coding efficiency beyond that obtained through intra prediction. Some of them include jointly encoding YUV coefficients that are equal to zero, and adaptive quantization schemes of AC coefficients. These techniques could also be applied to the RAGFT but their implementation goes beyond the scope of this paper and we leave them for future work. In order to obtain a performance comparison between our method and the current state-of-the-art, under the same conditions, we changed the source code of G-PCC modifying both the quantizer and the entropy encoder schemes. The adaptive quantizer was replaced with a uniform quantizer and the entropy encoder was modified to an adaptive RLGR algorithm. Moreover, we encode YUV coefficients independently. In Figure 4 we report rate distortion curves for the “longdress” and “loot” point clouds of the “8iVFBv2” dataset [23].
Incorporating intra prediction into the RAGFT improves coding performance significantly. The difference between “RAGFT ” and “I-RAGFT LowRes” is about db, while the gain obtained by using better predictors (“I-RAGFT Proposed”) can reach up to db. In [13], the RAGFT implemented with larger block sizes () led to the best results, since spatial redundancy can be removed more efficiently using graph transforms on larger blocks. Our results show that by combining low resolution intra prediction and RAGFT with small block size (“I-RAGFT ”), we can outperform the best RAGFT at low bitrates. However this approach is still inferior to I-RAHT. The proposed predictor based on the higher resolution point cloud, combined with the RAGFT achieves the best rate distortion performance among all methods considered, and outperforms I-RAHT by up to db.
In Figure 5 we compare decoded attributes obtained after compression using I-RAGFT, I-RAHT and RAGFT (), at the same rate of . Transform coding with the RAGFT produces strong blocking artifacts, similar to those associated to JPEG encoding. Both intra coding approaches have improved visual quality with respect to RAGFT, while also using smaller transforms, thus avoiding blocking artifacts. I-RAHT produces an over smoothed reconstruction, which can be accredited to the less localized filtering operations. Since the I-RAGFT uses higher resolution predictors, image details are better preserved, however it suffers from other localized artifacts.
5 Conclusion
We studied the use of intra prediction for compression of 3D point cloud attributes with the RAGFT. As with the RAHT, we showed that it is possible to remove redundancy between RAGFT coefficients at different resolutions. In the RAHT, a high resolution point cloud is predicted from a lower resolution point cloud. While this approach is efficient for predicting RAHT coefficients, it is less effective for the RAGFT. This is because the RAGFT uses larger block transforms, that makes prediction more challenging. To overcome this issue, we proposed a different prediction based on interpolation and graph signal filtering, that takes into account the higher resolution geometry, thus adapting better to the RAGFT. Compression experiments show the proposed approach outperforms intra predictive coding of RAHT coefficients at all rates by up to db.
References
- [1] Sebastian Schwarz, Marius Preda, Vittorio Baroncini, Madhukar Budagavi, Pablo Cesar, Philip A Chou, Robert A Cohen, Maja Krivokuća, Sébastien Lasserre, Zhu Li, et al., “Emerging MPEG standards for point cloud compression,” IEEE Journal on Emerging and Selected Topics in Circuits and Systems, vol. 9, no. 1, pp. 133–148, 2018.
- [2] D Graziosi, O Nakagami, S Kuma, A Zaghetto, T Suzuki, and A Tabatabai, “An overview of ongoing point cloud compression standardization activities: Video-based (V-PCC) and geometry-based (G-PCC),” APSIPA Transactions on Signal and Information Processing, vol. 9, 2020.
- [3] Eduardo Pavez, Philip A Chou, Ricardo L De Queiroz, and Antonio Ortega, “Dynamic polygon clouds: representation and compression for VR/AR,” APSIPA Transactions on Signal and Information Processing, vol. 7, 2018.
- [4] Ricardo L De Queiroz and Philip A Chou, “Compression of 3D point clouds using a region-adaptive hierarchical transform,” IEEE Transactions on Image Processing, vol. 25, no. 8, pp. 3947–3956, 2016.
- [5] 3DG, “G-PCC codec description,” Tech. Rep. W19620, ISO/IEC JTC1/SC29/WG11 MPEG, 2020.
- [6] Yiqun Xu, Wei Hu, Shanshe Wang, Xinfeng Zhang, Shiqi Wang, Siwei Ma, Zongming Guo, and Wen Gao, “Predictive generalized graph Fourier transform for attribute compression of dynamic point clouds,” IEEE Transactions on Circuits and Systems for Video Technology, 2020.
- [7] André L Souto and Ricardo L de Queiroz, “On predictive RAHT for dynamic point cloud coding,” in 2020 IEEE International Conference on Image Processing (ICIP). IEEE, 2020, pp. 2701–2705.
- [8] Ricardo L de Queiroz and Philip A Chou, “Motion-compensated compression of dynamic voxelized point clouds,” IEEE Transactions on Image Processing, vol. 26, no. 8, pp. 3886–3895, 2017.
- [9] Dorina Thanou, Philip A Chou, and Pascal Frossard, “Graph-based compression of dynamic 3D point cloud sequences,” IEEE Transactions on Image Processing, vol. 25, no. 4, pp. 1765–1778, 2016.
- [10] Robert A Cohen, Dong Tian, and Anthony Vetro, “Point cloud attribute compression using 3-d intra prediction and shape-adaptive transforms,” in 2016 Data Compression Conference (DCC). IEEE, 2016, pp. 141–150.
- [11] Yiting Shao, Qi Zhang, Ge Li, Zhu Li, and Li Li, “Hybrid point cloud attribute compression using slice-based layered structure and block-based intra prediction,” in Proceedings of the 26th ACM international conference on Multimedia, 2018, pp. 1199–1207.
- [12] Sébastien Lasserre and David Flynn, “On an improvement of RAHT to exploit attribute correlation,” input document m47378, ISO/IEC JTC1/SC29/WG11 MPEG, 2019.
- [13] Eduardo Pavez, Benjamin Girault, Antonio Ortega, and Philip A. Chou, “Region adaptive graph Fourier transform for 3D point clouds,” in International Conference on Image Processing (ICIP), 2020.
- [14] Antonio Ortega, Pascal Frossard, Jelena Kovačević, José MF Moura, and Pierre Vandergheynst, “Graph signal processing: Overview, challenges, and applications,” Proceedings of the IEEE, vol. 106, no. 5, pp. 808–828, 2018.
- [15] Peyman Milanfar, “A tour of modern image filtering: New insights and methods, both practical and theoretical,” IEEE signal processing magazine, vol. 30, no. 1, pp. 106–128, 2012.
- [16] Gene Cheung, Enrico Magli, Yuichi Tanaka, and Michael K Ng, “Graph spectral image processing,” Proceedings of the IEEE, vol. 106, no. 5, pp. 907–930, 2018.
- [17] B. Girault, A. Ortega, and S.S. Narayanan, “Irregularity-aware graph Fourier transforms,” IEEE Transactions on Signal Processing, vol. 66, no. 21, pp. 5746–5761, 2018.
- [18] Antoni Dimitriadis and David Taubman, “Augmenting JPEG2000 with wavelet coefficient prediction,” in 2020 IEEE International Conference on Image Processing (ICIP). IEEE, 2020, pp. 1276–1280.
- [19] Henrique S Malvar, “Adaptive run-length/Golomb-Rice encoding of quantized generalized gaussian sources with unknown statistics,” in Data Compression Conference (DCC’06). IEEE, 2006, pp. 23–32.
- [20] Cha Zhang, Dinei Florencio, and Charles Loop, “Point cloud attribute compression with graph transform,” in 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014, pp. 2066–2070.
- [21] C.L. Jackins and S. L. Tanimoto, “Oct-trees and their use in representing three-dimensional objects,” Computer Graphics and Image Processing, vol. 14, no. 3, pp. 249–270, 1980.
- [22] D. Meagher, “Geometric modeling using octree encoding,” Computer graphics and image processing, vol. 19, no. 2, pp. 129–147, 1982.
- [23] Eugene d’Eon, Bob Harrison, Taos Myers, and Philip A. Chou, “8i voxelized full bodies-a voxelized point cloud dataset,” ISO/IEC JTC1/SC29 Joint WG11/WG1 (MPEG/JPEG) input document WG11M40059/WG1M74006, 2017.