ArticlePDF Available

POINT-CLOUD COMPRESSION FOR VEHICLE-BASED MOBILE MAPPING SYSTEMS USING PORTABLE NETWORK GRAPHICS

Authors:

Abstract and Figures

A mobile mapping system is effective for capturing dense point-clouds of roads and roadside objects.Point-clouds of urban areas, residential areas, and arterial roads are useful for maintenance of infrastructure, map creation, and automatic driving. However, the data size of point-clouds measured in large areas is enormously large. A large storage capacity is required to store such point-clouds, and heavy loads will be taken on network if point-clouds are transferred through the network. Therefore, it is desirable to reduce data sizes of point-clouds without deterioration of quality. In this research, we propose a novel point-cloud compression method for vehicle-based mobile mapping systems. In our compression method, point-clouds are mapped onto 2D pixels using GPS time and the parameters of the laser scanner. Then, the images are encoded in the Portable Networking Graphics (PNG) format and compressed using the PNG algorithm. In our experiments, our method could efficiently compress point-clouds without deteriorating the quality.
Content may be subject to copyright.
POINT-CLOUD COMPRESSION FOR VEHICLE-BASED MOBILE MAPPING SYSTEMS
USING PORTABLE NETWORK GRAPHICS
K. Kohira a, H. Masuda a, *
a Dept. of Mechanical Engineering and Intelligent Systems, The University of Electro-Communications,
1-5-1 Chofugaoka, Chofu, Tokyo, Japan - (keisuke.kohira, h.masuda)@uec.ac.jp
Commission VI, WG VI/4
KEY WORDS: Mobile Mapping System, point-clouds, lossless compression
ABSTRACT:
A mobile mapping system is effective for capturing dense point-clouds of roads and roadside objectsPoint-clouds of urban areas,
residential areas, and arterial roads are useful for maintenance of infrastructure, map creation, and automatic driving. However, the
data size of point-clouds measured in large areas is enormously large. A large storage capacity is required to store such point-clouds,
and heavy loads will be taken on network if point-clouds are transferred through the network. Therefore, it is desirable to reduce data
sizes of point-clouds without deterioration of quality. In this research, we propose a novel point-cloud compression method for
vehicle-based mobile mapping systems. In our compression method, point-clouds are mapped onto 2D pixels using GPS time and the
parameters of the laser scanner. Then, the images are encoded in the Portable Networking Graphics (PNG) format and compressed
using the PNG algorithm. In our experiments, our method could efficiently compress point-clouds without deteriorating the quality.
1. INTRODUCTION
The vehicle-based mobile mapping system (MMS) is effective
for capturing 3D shapes of roads, buildings, and roadside
objects. Figure 1 shows an MMS, on which laser scanners,
cameras, GPSs and IMU are mounted. The MMS captures
point-clouds using laser scanners while running. Point-clouds of
urban areas, residential areas, and arterial roads are useful for
maintenance of infrastructure, map creation, automatic driving,
computer graphics, and so on.
When high-performance laser scanners are installed on the
MMS, the data size of point-clouds becomes very enormous. In
recent years, the performance of laser scanners has been greatly
improved. For example, the maximum data acquisition rate of
the Z+F Profiler 9012 is 1.016 million measurements per
second, and the one of the RIEGL VQ 450 is 0.55 million
measurements per second. When the MMS with a high-
performance laser scanner runs over a long distance, billions of
points are captured in one day. While dense point-clouds
maintain rich 3D information, the data volume often becomes
extraordinary large. Many hard disks are required to store large-
scale data, and it takes a long time to load the data from hard
disks on RAM.
In order to reduce the storage capacity and shorten loading time
for big data, data compression techniques are very important.
Data compression reduces data sizes by encoding measured data
in fewer bits.
In laser scanning, effective digits of 3D coordinates can be
reduced according to the measurement accuracy of the MMS. In
most cases, it is sufficient to record 3D coordinates up to 1 mm.
However, even if the number of significant digits is reduced, the
data size is still very large. Therefore, it is necessary to further
reduce the data size of point-clouds using coherency among
coordinates. 1
* Corresponding author
Several point-cloud compression methods have been proposed
so far. There are three typical types of compression methods for
point-clouds.
The first type of method is based on coherency between
consecutive points. LASZIP proposed by (Isenburg, 2013) has
been widely used to compress LiDAR data. This method
compressed organized points based on the observation that
coordinates of consecutive points are similar. However, it does
not use adjacency relationships on 2D space in cases of MMS
data. (He, et al. 2012) also compressed MMS data by encoding
consecutive points as second order differences.
The second type of method is space partition for unorganized
points. Most space partition methods for point-clouds are based
on the octree or the k-d tree. The both methods recursively
subdivide three-dimensional space and creates hierarchical
structure of points. In octree-based methods, (Peng et al., 2003)
and (Huang et al., 2006) proposed progressive lossless mesh
encoders. (Kammerl et al., 2012) compressed unordered point-
cloud streams by encoding differences between voxels.
(Elseberg et al., 2013) proposed octree-based data structure to
Figure 1. Mobile mapping system
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
99
handle large points captured by terrestrial laser scanners.
(Hornung et al., 2013) proposed an octree map compression
method using probabilistic occupancy estimation. In kd-tree
based methods, (Gandoin et al., 2002) proposed a progressive
connectivity-coding algorithm using the kd-tree geometric coder
proposed by (Devillers et al., 2000). (Hubo et al., 2006)
proposed the Quantized kd-tree, in which points were encoded
using quantized split-plane positions.
The third type of method is based on 2D images for organized
points. If 3D points can be mapped on a 2D lattice image, point-
clouds would be more efficiently encoded. In terrestrial laser
scanning, it is well-known that a point-cloud can be mapped on
a 2D panoramic image if the scanner position is fixed. Even in
mobile mapping, some researchers converted points into images
for specific laser scanners. (Houshiar, 2015) mapped point-
clouds captured at the fixed positions onto panoramic depth
images, and encoded depth images using an image compression
method. For MMS data, (Kaess et al., 2003) proposed a
compression method for laser scanners that emit laser beams
line by line, such as the SICK LMS series. They aligned scan
lines in a lattice manner, and compressed them using an image
compression method. (Masuda et al., 2015) also proposed a
scan-line based compression method, in which a sequence of
points was encoded using the second-order differences. (Tu et
al., 2016) compressed raw point data of the Velodyne HDL-
64S2, which emits 64 laser beams simultaneously. They
converted a set of 64 scan lines into a panoramic image and
applied compression methods for images and videos. However,
these methods for mobile mapping data were developed only for
specific types of laser scanners. A majority of laser scanners for
MMSs, such as RIEGL VX and Z+F Profiler series,
continuously emit laser beams in a spiral manner, as shown in
Figure 2. In conventional methods, it is difficult to efficiently
compress MMS data captured by these popular laser scanners.
In this paper, we propose a novel image-based compression
method for point-clouds captured by MMSs. Our main
contribution is a novel compression framework for spiral-type
laser scanners, and a novel image-based compression method
for MMS data. In our method, point-clouds are projected onto a
2D lattice using GPS times and laser scanner parameters. Then,
we segment point-clouds on the 2D lattice into groups of
neighbor points, and compress each group of points using the
Portable Network Graphics (PNG). Our method can compress
point-clouds captured by MMSs without deteriorating the
quality.
2. MAPPING POINT-CLOUDS ONTO 2D IMAGES
2.1 Mapping 3D Points on 2D Lattice
Point-clouds captured by MMSs do not maintain explicit 2D
lattice structure, while point-clouds captured by TLSs can be
easily converted to panoramic images. Therefore, we first
consider a method for mapping point-clouds onto 2D images by
assigning the pulse number and the rotation number to each
point.
In spiral-type laser scanners, laser beams are continuously
emitted, as shown in Figure 2. Points are measured at the equal
time interval. We suppose that point-clouds include (𝑥,𝑦,𝑧)
coordinates, intensity values, and GPS times. A GPS time
indicates when the point was captured. RGB colors are often
added to points in post processing, but we do not discuss RGB
colour attributes in this paper, because compression of RGB
colours is trivial. Since point-clouds are converted into images
in our method, RGB colours can be simply compressed using
any image compression method.
In this paper, let {𝑝!} be the sequence of points, 𝑡𝑖 be the
elapsed time from the start of laser scanning until point 𝑝𝑖 is
measured. We also denote the rotation frequency of the laser
scanner as 𝑓, and the pulse repetition frequency of the laser
scanner as 𝜔. The rotation frequency is the number of rotations
Figure 2. Trajectory of the laser beams
Figure 3. Neighbour points on scan-lines
Figure 4. The rotation number
Figure 5. The phase number
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
100
of laser beams per second, and the pulse repetition frequency is
the number of measurements per 1/𝑓 second.
Since the laser beam rotates once in 1/𝑓 second, we subdivide a
point-cloud every 1/𝑓 second. We call each segment as a scan-
line. As shown in Figure 3, neighbour points on the next scan-
line are captured 1/𝑓 second later.
In order to map a point-cloud on a 2D image, we assign the
phase number and the rotation number to each point. The
rotation number is the sequential number of scan-lines, as
shown in Figure 4. Points on the same scan-line have the same
rotation number. The rotation number of point pi can be
calculated as:
𝐽!=INT(𝑓𝑡!)
(1)
INT(𝑥) is the function that returns the integer part of 𝑥.
The phase number 𝐼𝑖 indicates the sequential order of point p!
on the scan-line. Figure 5 shows the phase numbers. Points with
the same phase number are ordered on the same column. The
phase number can be calculated as:
𝐼!=INT(𝜔FMOD 𝑡!,1/𝑓)
(2)
FMOD(𝑥,𝑦) is the function that returns the floating-point
remainder of 𝑥/𝑦.
Each point pi is mapped to (𝐼!,𝐽!) on the 2D lattice. If any point
is not mapped on a pixel, the pixel is marked as empty. Figure 6
shows an example of mapped points. In this figure, the vertical
axis is the phase number, and the horizontal axis is the rotation
number. The brightness of a pixel is the intensity value of each
point. Empty pixels are shown in blue colour.
2.2 Quantization of Coordinates
Coordinates measured by MMS include numerical errors caused
by laser scanners, GPS, calibration, and so on. Since the
accuracy of coordinate values is not very high, very high
resolution, such as sub millimeter, is not necessary.
To reduce the number of coordinate digits, we quantize point
coordinates according to the accuracy of the MMSs. In this
research, we set the quantization step as 1mm. This resolution is
sufficiently high considering the accuracy of MMS data.
2.3 Segmentation of 2D Lattice
Although we reduce the number of digits by quantizing the
coordinates, the number of digits is still very large. Therefore,
we represent coordinates as differences between neighbors to
make the data more compressible. In this paper, we consider
dividing the image so that the differences of 𝑥, 𝑦, and 𝑧 can be
encoded with 8 bits.
We segment the image into connected regions using the region
growing method. The point with the earliest GPS time is
selected as a seed point. Initially, the region contains only the
seed region. Then, neighbour points are added to the same
region only when the difference for each of 𝑥,𝑦,𝑧 is less than
256 mm. This process is repeated until neighbour points cannot
be added anymore. When there are unsegmented points, a new
seed point is selected from the remaining points, and the region
growing method is applied again.
Figure 7 shows the segmented image, in which connected
regions are shown in different colours. Figure 8 shows images
of connected regions. In our method, each region is converted
into the Portable Network Graphics (PNG) format.
3. POINT-CLOUD COMPRESSION USING
PORTABLE NETWORK GRAPHICS
3.1 Portable Network Graphics
PNG (W3C, 2003) is a patent free file format designed for
images. PNG supports lossless data compression, which can
completely reconstruct the original image from the compressed
data. The PNG algorithm consists of filtering and compression
stages. In the filtering stage, images are passed through delta
Figure 6. 2D lattice projected point-clouds
Figure 7. Connected regions
Figure 8. Images of connected regions
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
101
filtering to make the image compressible. In the compression
stage, Deflate algorithm (Deutsch, 1996) is applied to the
filtered byte sequences. In this compression algorithm, LZ77
cording (Ziv et al., 1977) and Huffman cording are combined.
The PNG format supports several types of images. We use the
PNG format for RGB images with alpha channels to encode
point-clouds.
3.2 Differences of Coordinates
When connected regions are detected using the region growing
method, each pixel is added one by one. Figure 9 shows the
process of growing region. In this figure, when pixel 𝑃
! is
added to the connected region as the neighbor of pixel 𝑃
!, a
directed edge is defined from pixel 𝑃
! to pixel 𝑃
!. Since the all
nodes in the connected region are traversed during the region
growing, this graph becomes a directed spanning tree. When the
seed point is given, the spanning tree is uniquely reconstructed
according to the rule of region growing.
Once the directed spanning tree is generated, coordinate
differences can be calculated between the adjacent nodes in the
directed graph. As shown in Figure 10, the coordinate
difference 𝑝! is calculated between the seed and the neighbor
point on the spanning tree. Similarly, differences
𝑝!,𝑝!,,𝑝! are calculated along the spanning tree. The
difference value is stored at each pixel. Difference values are
guaranteed to be 255 mm or less based on the rule of region
growing.
Coordinate 𝑝! can be reconstructed using the seed coordinate
𝑝!""# and difference values. Let Λ! be the index set of the
shortest path on the spanning tree from the seed to pixel 𝑖. Then
𝑝! can be calculated as:
𝑝!=𝑝!""# +
!!!
𝑝!
(3)
Therefore, in our method, the coordinate of the seed point and
the differences of other pixels are required to reconstruct all
original coordinates.
3.3 Encoding Point-Cloud
In our method, difference Δ𝑥!,Δ𝑦!,Δ𝑧! is stored at pixel 𝑖. We
encode them in the PNG format, which has RGB channels and
an alpha channel. In this format, 8 bits are allocated to each
channel of R, G, B, and alpha.
Figure 11 shows our encoding scheme. We as si gn three bits to
plus or minus signs of (𝑥,𝑦,𝑧), and store them in the alpha
channel as 0 or 1. The absolute values of (𝑥,𝑦,𝑧) are written in
RGB channels using 8 bits. The remaining 5 bits in the alpha
channel can be used optionally. In this paper, we use the 5 bits
for storing an intensity value, which is quantized to 25 = 32
levels.
The coordinate data in the PNG format is compressed and
stored in the file with .png extension. Figure 12 shows an
intensity image of a connected region and its PNG image. In
Figure 12(b), pixel colors are determined from coordinate
differences. We can reconstruct point coordinates and intensity
values from the encoded image. Figure 13 shows a point-cloud
reconstructed from a PNG image. Figure 13(a) encodes a point-
cloud equivalent to Figure 13(b).
The GPS time of each point can be also reconstructed, although
GPS times are not explicitly encoded in the file. In
measurement by a MMS, points are measured at the equal time
interval, and they are sequentially ordered on the image.
Therefore, the GPS time of each point can be determined using
the position on the image if the earliest GPS time of the seed
points is given. The GPS time 𝑇
! at pixel (𝐼!,𝐽!) can be
calculated using the GPS time 𝑇
!""# of the seed point as:
𝑇
!=𝑇
!""# +
𝐼!+𝜔𝐽!
𝜔𝑓
(4)
Figure 9. Spanning tree based on region growing
Figure 10. Difference values
Figure 11. Encoding coordinate difference
(a) Intensity image
(b) PNG image
Figure 12. Encoded image of a point-cloud
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
102
In our method, the coordinate and GPS time of the seed are
required to reconstruct the original coordinates and GPS times.
In our implementation, these values are stored as the file name
of compressed data.
3.4 Encoding Fragmented Regions
In our method, each connected region is compressed as a PNG
file. When an image is segmented into connected regions, large
regions are generated from buildings, roads, and roadside
objects. On the other hand, small fragmented regions are also
generated from tree leaves, cables, and noises. In Figure 14,
while a large connected region is generated from a building,
many fragmented regions are generated from an electric cable.
In our method, while the compression efficiency for large
regions is very high, the efficiency is poor for fragmented
regions with few points. Therefore, we compress only large
regions by the PNG algorithm. In this paper, we regard regions
with 5 or less points as fragmented regions.
In Figure 15, fragmented regions are shown in red colour. In
residential areas, fragmented regions are mainly generated from
cables and trees.
Table 1 shows ratios of points in fragmented regions for point-
clouds A and B in Figure 16. Since many trees are included in
the point-cloud B, more fragmented regions are generated from
tree leaves. In both cases, the numbers of points in fragmented
regions are very small, because areas of tree leaves and cables
are much smaller than ones of roads and building walls,
In our method, points in fragmented regions are separately
encoded without using the PNG algorithm. In this paper, we
simply encode points in fragmented regions in a binary format
without compression, because the number of points in
fragmented regions is relatively very small.
4. EXPERIMANTAL RESULTS
We evaluated our compression method using two datasets A
and B, which were measured in a residential district in Japan
using Mitsubishi Electric MMS Type X. The mounted laser
scanner was Z+F Profiler 9012. As shown in Figure 16, point-
clouds A and B include houses, roads, utility poles, and so on.
Point-cloud B also includes many trees, and therefore a lot of
fragmented regions are generated from dataset B. The numbers
of points are 9.79 million in point-cloud A, and 6.54 million in
point-cloud B. For data compression, we used a PC with 3.1
GHz Intel core i5 CPU and 16 GB RAM.
We compared our method with the scan-line based compression
ASCII
Binary
Scan-Line
Based
Our
Method
Size
552.4 MB
319.5 MB
30.1 MB
23.0 MB
Ratio
100 %
57.8 %
5.4 %
4.2 %
Table 2. Compression results of point-cloud A
ASCII
Binary
Scan-Line
Based
Our
Method
Size
367.6 MB
213.2 MB
24.1 MB
18.4 MB
Ratio
100 %
58.0 %
6.6 %
5.0 %
Table 3. Compression results of point-cloud B
(a) PNG image
(b) Points
Figure 13. Image and points representing the same object
(a) Large region
(b) Fragmented region
Figure 14. Sizes of connected regions
(a) Electric cable (b) Tree leaves
Figure 15. Fragmented regions
Number of Points
Large Regions
Fragmented Region
Data A
9,762,923 (99.7%)
30,365 (0.3%)
Data B
6,448,844 (98.7%)
86,445 (1.3%)
Table 1. Number of points in large and fragmented regions
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
103
method (Masuda et al., 2015). This method was designed for
MMS data, and compresses scan-lines using the second-order
differences. In our evaluation, coordinates, GPS times, and
intensity values were encoded in a file. The resolution of 𝑥, 𝑦,
and 𝑧 was 1 mm, and the intensity values were stored in 5 bits.
The results are shown in Table 2 and Table 3. In ASCII and
binary formats, data were stored without compression. In the
both datasets, our method has improved compression ratios by
more than 20% compared to the scan-line based compression
method.
Our method showed excellent compression ratios even when
many fragmented regions were generated. In our method, while
large regions are encoded using the PNG algorithm, fragmented
regions are stored without compression. In Table 4, data sizes
for both regions are described. In these datasets, the
fragmentation problem did not have significant effect on data
compression.
Figure 17 shows point-clouds that were reconstructed from
compressed data. Fragmented points are shown in red colour.
Figure 18-20 show details of reconstructed point-clouds. In all
cases, the original point-clouds could be faithfully reconstructed,
because our compression method is lossless after coordinates
are quantized to 1 mm resolution.
After compressed data are loaded from a hard disk, it has to be
decoded for use. Table 5 shows timing for loading and decoding
of compressed data. We compared with loading time of original
point-clouds represented in the ASCII format. The time for
loading and decoding compressed data was shorter than the
loading time of the original point-cloud.
Table 6 shows timing for compression process. We measured
calculation time for 2D mapping, segmentation, encoding to the
PNG format, and saving as a PNG file. In this evaluation, most
(a) Point-cloud A
(b) Point-cloud B
Figure 16. Original point-clouds
(a) Reconstructed point-cloud A
(b) Reconstructed point-cloud B
Figure 17. Reconstructed point-clouds
Dataset
PNG image
Fragmented points
A
22.4 MB (97 %)
0.6 MB (3 %)
B
16.5 MB (89 %)
1.9 MB (11 %)
Table 4. Data sizes of PNG images and fragmented points
Dataset
Compressed Data
Original Data
Load
Decode
Total
Load
A
1.0 s
4.0 s
5.0 s
13.9 s
B
0.9 s
2.6 s
3.5 s
9.1 s
Table 5. Timing for loading and decoding
Dataset
2D
Mapping
Segmenta
tion
Encoding
Writing
PNG
A
2.4 s
4.0 s
12.8 s
1.0 s
B
1.6 s
2.9 s
7.7 s
2.0 s
Table 6. Timing for compression process
Dataset
ΔX
ΔY
ΔZ
A
0.53
0.50
0.47
B
0.54
0.49
0.47
Table 7. Average coordinate errors (mm)
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
104
computation time was used for segmentation and encoding
processes. The total computation time is 20.2 sec for point-
cloud A, and 14.2 sec for point-cloud B.
In our method, coordinates are quantized to 1 mm resolution.
Table 7 shows the average distance errors between the original
points without quantization and the reconstructed points. In the
both datasets, the average distance errors were about 0.5 mm.
Considering the measurement accuracy of MMSs, the resolution
of reconstructed point-clouds are sufficient.
5. CONCLUSION
We proposed a novel compression method for point-clouds
captured by a vehicle-based mobile mapping system. In our
method, point-clouds were projected onto pixels of an image
using the rotation number and the phase number, and the image
was segmented into connected regions. Then points in each
connected region were encoded to the PNG format, and
compressed using the PNG algorithm. In our experiments, our
method could achieve better compression ratios compared to the
scan-based compression method.
In future work, we would like to improve compression rates.
Since there are other lossless compression methods, such as
JPEG 2000, we would like to investigate other compression
schemes. In our method, it takes time to encode point-clouds.
We would like to improve our algorithm to shorten encoding
time. Currently, we evaluated our method only in residential
areas. We would like to evaluate our method using point-clouds
of highways and suburbs.
ACKNOWLEDGEMENTS
MMS data in this paper are courtesy of AISAN Technology
Co.Ltd. We would like to thank for their helpful support.
REFERENCES
Devillers, O., Gandoin, P. M., 2000. Geometric compression for
interactive transmission. IEEE Visualization 2000. Proceedings,
pp. 319-326.
Deutsch, L. P., 1996. DEFLATE compressed data format
specification version 1.3. RFC Editor.
Elseberg, J., Borrmann, D., Nüchter, A., 2013. One billion
points in the cloudan octree for efficient processing of 3D laser
scans. ISPRS Journal of Photogrammetry and Remote
Sensing, 76, pp. 76-88.
Gandoin, P. M., Devillers, O., 2002. Progressive lossless
compression of arbitrary simplicial complexes. ACM
Transactions on Graphics (TOG), 21(3), pp.372-379.
He J., Masuda H., 2012. Reconstruction of Roadways and
Walkways Using Point-Clouds from Mobile Mapping System,
Asian Conference on Design and Digital Engineering, Article
100099.
Hornung, A., Wurm, K. M., Bennewitz, M., Stachniss, C.,
Burgard, W., 2013. OctoMap: An efficient probabilistic 3D
mapping framework based on octrees. Autonomous Robots,
34(3), pp. 189-206.
(a) Original points
(b) Reconstructed points
Figure 18. Points of roads and a car
(a) Original points
(b) Reconstructed points
Figure 19. Points of a traffic sign
(a) Original points
(b) Reconstructed points
Figure 20. Points of electric cables
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
105
Houshiar, H., Nüchter, A., 2015. 3D point cloud compression
using conventional image compression for efficient data
transmission. Information, Communication and Automation
Technologies (ICAT), 2015 XXV International Conference, pp.
1-8.
Huang, Y., Peng, J., Kuo, C. C. J., Gopi, M. 2006. Octree-Based
Progressive Geometry Coding of Point Clouds. Symposium on
Point-Based Graphics pp. 103-110.
Hubo, E., Mertens, T., Haber, T., Bekaert, P., 2006. The
quantized kd-tree: Efficient ray tracing of compressed point
clouds. Interactive Ray Tracing 2006, IEEE Symposium, pp.
105-113.
Isenburg, Martin., 2013. Laszip. Photogrammetric Engineering
& Remote Sensing 79(2), pp.209-217.
Masuda, H., He, J., 2015. TIN generation and point-cloud
compression for vehicle-based mobile mapping systems.
Advanced Engineering Informatics, 29(4), pp.841-850.
Kaess, M., Arkin, R. C., Rossignac, J., 2003. Compact encoding
of robot-generated 3D maps for efficient wireless transmission.
IEEE Intl. Conf. on Advanced Robotics (ICAR), pp. 324-331.
Kammerl, J., Blodow, N., Rusu, R. B., Gedikli, S., Beetz, M.,
Steinbach, E., 2012. Real-time compression of point cloud
streams. Robotics and Automation (ICRA), 2012 IEEE
International Conference, pp. 778-785.
Peng, J., Kuo, C. J., 2003. Octree-based progressive geometry
encoder. ITCom 2003. International Society for Optics and
Photonics, pp. 301-311.
Tu, C., Takeuchi, E., Miyajima, C., Takeda, K., 2016.
Compressing continuous point cloud data using image
compression methods. Intelligent Transportation Systems
(ITSC), 2016 IEEE 19th International Conference, pp. 1712-
1719.
W3C., 2003. Portable network graphics (png) specification,
http://www.w3.org/TR/PNG/.
Ziv, J., Lempel, A., 1977. A universal algorithm for sequential
data compression. IEEE Transactions on information
theory, 23(3), pp.337-343.
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume IV-2/W4, 2017
ISPRS Geospatial Week 2017, 18–22 September 2017, Wuhan, China
This contribution has been peer-reviewed. The double-blind peer-review was conducted on the basis of the full paper.
https://doi.org/10.5194/isprs-annals-IV-2-W4-99-2017 | © Authors 2017. CC BY 4.0 License.
106
... In [22] and [23] LiDAR data are flattened into 2D range images that are then compressed using image compression methods. A Simultaneous Localization And Mapping (SLAM) approach towards predicting and compress consecutively acquired PCs is presented in [24]. ...
Article
Full-text available
The rapid growth on the amount of generated 3D data, particularly in the form of Light Detection And Ranging (LiDAR) point clouds (PCs), poses very significant challenges in terms of data storage, transmission, and processing. Point cloud (PC) representation of 3D visual information has shown to be a very flexible format with many applications ranging from multimedia immersive communication to machine vision tasks in the robotics and autonomous driving domains. In this paper, we investigate the performance of four reference 3D object detection techniques, when the input PCs are compressed with varying levels of degradation. Compression is performed using two MPEG standard coders based on 2D projections and octree decomposition, as well as two coding methods based on Deep Learning (DL). For the DL coding methods, we used a Joint Photographic Experts Group (JPEG) reference PC coder, that we adapted to accept LiDAR PCs in both Cartesian and cylindrical coordinate systems. The detection performance of the four reference 3D object detection methods was evaluated using both pre-trained models and models specifically trained using degraded PCs reconstructed from compressed representations. It is shown that LiDAR PCs can be compressed down to 6 bits per point with no significant degradation on the object detection precision. Furthermore, employing specifically trained detection models improves the detection capabilities even at compression rates as low as 2 bits per point. These results show that LiDAR PCs can be coded to enable efficient storage and transmission, without significant object detection performance loss.
... In [11], compression was attained through the use of Random Sample Consensus (RANSAC) to extract planes from a point cloud with the planes subsequently transformed into surfaces by Delaunay triangulation. There are also 2D representation methods, which alternatively propose mapping the point cloud into panoramic images [12] or, using additional laser scanner information, into 2D pixels [13]. A deep learning approach to compression can be seen in [14], where the authors propose employing a recurrent neural network to compress LiDAR data that had previously been converted to 2D images, while using a network with a residual block to decompress it. ...
Article
Full-text available
Increasing demand for more reliable and safe autonomous driving means that data involved in the various aspects of perception, such as object detection, will become more granular as the number and resolution of sensors progress. Using these data for on-the-fly object detection causes problems related to the computational complexity of onboard processing in autonomous vehicles, leading to a desire to offload computation to roadside infrastructure using vehicle-to-infrastructure communication links. The need to transmit sensor data also arises in the context of vehicle fleets exchanging sensor data, over vehicle-to-vehicle communication links. Some types of sensor data modalities, such as Light Detection and Ranging (LiDAR) point clouds, are so voluminous that their transmission is impractical without data compression. With most emerging autonomous driving implementations being anchored on point cloud data, we propose to evaluate the impact of point cloud compression on object detection. To that end, two different object detection architectures are evaluated using point clouds from the KITTI object dataset: raw point clouds and point clouds compressed with a state-of-the-art encoder and three different compression levels. The analysis is extended to the impact of compression on depth maps generated from images projected from the point clouds, with two conversion methods tested. Results show that low-to-medium levels of compression do not have a major impact on object detection performance, especially for larger objects. Results also show that the impact of point cloud compression is lower when detecting objects using depth maps, placing this particular method of point cloud data representation on a competitive footing compared to raw point cloud data.
... Compared to the traditional static mapping methods, mobile mapping is more efficient and flexible in different environments [1] and has become a rapidly evolving research field. Due to these advantages, it has been widely adopted to collect geospatial data in the fields of smart cities and autonomous driving [2]. Apart from efficiency and flexibility, data accuracy is also a significant piece of the puzzle in the mobile mapping system (MMS). ...
Article
Full-text available
A mobile mapping system (MMS) is a widely-used moving platform for the acquisition of rich geospatial information and requires sensor calibration to guarantee data quality. Global navigation satellite system (GNSS) sensor and light detection and ranging (LiDAR) are important parts of an MMS, and the extrinsic calibration between them can markedly improve its performance. A sole GNSS sensor is generally used in a portable MMS, which is the current trend. However, existing extrinsic calibration methods of GNSS/inertial navigation system (INS) are not applicable for the sole GNSS sensor because the former requires relative six degrees of freedom (DoF) pose output for both sensors and the latter can only provide 3-DoF positions. Here, we propose two sole GNSS sensor extrinsic calibration methods: one is a target-based direct calculation method and the other is a non-target-based automatic iterative method. The experiment results show the two methods have good repeatability, the error of the target-based method is below 3 cm, and the error of the non-target-based method is below 2 cm.
... Compared to a traditional terrestrial laser scanner which is statically mounted on a tripod, it is more efficient and flexible, and can better adapt to complex environments [1]. Due to these advantages, it has become a hot research field and has been rapidly applied to the industries of land surveying, digital city, and autonomous driving [2]. ...
Article
Full-text available
A mobile mapping system (MMS) is a widely-used platform to collect geospatial information. However, in the monotonous environment, current methods have inadequate performance in mapping accuracy because of lacking geometric features and constraints for point cloud alignment. We propose a systematic pedestrian dead reckoning (PDR) augmentation mapping framework for backpack MMS. The framework starts with data acquisition, followed by our proposed lightweight monotonous scene recognition method based on statistical features. An indicator is also proposed to measure monotonous degrees. Then, a step detection of PDR based on four-layer long short-term memory (LSTM) networks is implemented. Lastly, the PDR information is fused with the LiDAR odometry by a factor graph (FG). Experiments are conducted in two common monotonous environments, a tunnel and a long narrow alley. The results show that adding the PDR information can improve the mapping accuracy from meter-level to decimeter-lever or even centimeter-level in less serious monotonous conditions.
... It is obvious that our method outperforms LOAM, one of the best LiDAR odometry methods, using as low as 3.36% LiDAR frames, and needs as low as 0.01% points for the feature point map and 0.30% points for the long-lasting map. In comparison, the existing point cloud compression methods [64], [40], [65], [66], [67], [68], [69], [70] require at least 2% points. As the distance between selected LiDAR frames enlarges, the map generation module uses fewer LiDAR frames to build the map, resulting in smaller map sizes and worse performance in localization. ...
Preprint
LiDAR odometry and localization has attracted increasing research interest in recent years. In the existing works, iterative closest point (ICP) is widely used since it is precise and efficient. Due to its non-convexity and its local iterative strategy, however, ICP-based method easily falls into local optima, which in turn calls for a precise initialization. In this paper, we propose CoFi, a Coarse-to-Fine ICP algorithm for LiDAR localization. Specifically, the proposed algorithm down-samples the input point sets under multiple voxel resolution, and gradually refines the transformation from the coarse point sets to the fine-grained point sets. In addition, we propose a map based LiDAR localization algorithm that extracts semantic feature points from the LiDAR frames and apply CoFi to estimate the pose on an efficient point cloud map. With the help of the Cylinder3D algorithm for LiDAR scan semantic segmentation, the proposed CoFi localization algorithm demonstrates the state-of-the-art performance on the KITTI odometry benchmark, with significant improvement over the literature.
Article
Vehicle localization is a critical component in autonomous driving systems, and LiDAR-based methods have become increasingly popular for this task. In this paper, we present a novel vehicle localization approach based on the point cloud map generated from LiDAR data. In particular, our approach first uses semantic segmentation and feature point extraction techniques to create an efficient feature point map and a long-lasting map from LiDAR sequences with corresponding poses. We then introduce a map-based online localization method that achieves precise vehicle localization using both LiDAR scans and the two point cloud maps, along with a multi-resolution ICP strategy. Comprehensive evaluations are conducted on the KITTI odometry dataset and the collected results demonstrate superior performance over the existing literature in both odometry metrics and absolute translation error. Our multi-resolution ICP-based method holds significant potential for map-based vehicle localization, offering promising prospects for application in autonomous driving and associated domains.
Article
For road mapping, it is important to add labels to point clouds captured by the Mobile Mapping System (MMS). Some automatic labeling methods have been proposed so far. However, in our experiment, conventional labeling methods were not sufficiently accurate for actual point clouds measured in Japan. In this paper, we propose a high-performance classification method that combines the multi-scale features of point clouds, the MMS specific features and the features obtained from point clouds mapped on the 2D image. The accuracy of the proposed method was evaluated using actual MMS data, and it was confirmed that the proposed method could achieve high recognition rate generalization performance. Our method can improve the accuracy of automatic labeling of point clouds, and is expected to improve the efficiency of map maintenance, which is a social infrastructure.
Article
The increase in popularity of point-cloud-oriented applications has triggered the development of specialized compression algorithms. In this paper, a novel algorithm is developed for the lossless geometry compression of voxelized point clouds following an intra-frame design. The encoded voxels are arranged into runs and are encoded through a single-pass application directly on the voxel domain. This is done without representing the point cloud via an octree nor rendering the voxel space through an occupancy matrix, therefore decreasing the memory requirements of the method. Each run is compressed using a context-adaptive arithmetic encoder yielding state-of-the-art compression results, with gains of up to 15% over TMC13 , MPEG’s standard for point cloud geometry compression. Several proposed contributions accelerate the calculations of each run’s probability limits prior to arithmetic encoding. As a result, the encoder attains a low computational complexity described by a linear relation to the number of occupied voxels leading to an average speedup of 1.8 over TMC13 in encoding speeds. Various experiments are conducted assessing the proposed algorithm’s state-of-the-art performance in terms of compression ratio and encoding speeds.
Thesis
With the rapid growth of multimedia content, 3D objects are becoming more and more popular. Most of the time, they are modeled as complex polygonal meshes or dense point clouds, providing immersive experiences in different industrial and consumer multimedia applications. The point cloud, which is easier to acquire than mesh and is widely applicable, has raised many interests in both the academic and commercial worlds.A point cloud is a set of points with different properties such as their geometrical locations and the associated attributes (e.g., color, material properties, etc.). The number of the points within a point cloud can range from a thousand, to constitute simple 3D objects, up to billions, to realistically represent complex 3D scenes. Such huge amounts of data bring great technological challenges in terms of transmission, processing, and storage of point clouds.In recent years, numerous research works focused their efforts on the compression of meshes, while less was addressed for point clouds. We have identified two main approaches in the literature: a purely geometric one based on octree decomposition, and a hybrid one based on both geometry and video coding. The first approach can provide accurate 3D geometry information but contains weak temporal consistency. The second one can efficiently remove the temporal redundancy yet a decrease of geometrical precision can be observed after the projection. Thus, the tradeoff between compression efficiency and accurate prediction needs to be optimized.We focused on exploring the temporal correlations between dynamic dense point clouds. We proposed different approaches to improve the compression performance of the MPEG (Moving Picture Experts Group) V-PCC (Video-based Point Cloud Compression) test model, which provides state-of-the-art compression on dynamic dense point clouds.First, an octree-based adaptive segmentation is proposed to cluster the points with different motion amplitudes into 3D cubes. Then, motion estimation is applied to these cubes using affine transformation. Gains in terms of rate-distortion (RD) performance have been observed in sequences with relatively low motion amplitudes. However, the cost of building an octree for the dense point cloud remains expensive while the resulting octree structures contain poor temporal consistency for the sequences with higher motion amplitudes.An anatomical structure is then proposed to model the motion of the point clouds representing humanoids more inherently. With the help of 2D pose estimation tools, the motion is estimated from 14 anatomical segments using affine transformation.Moreover, we propose a novel solution for color prediction and discuss the residual coding from prediction. It is shown that instead of encoding redundant texture information, it is more valuable to code the residuals, which leads to a better RD performance.Although our contributions have improved the performances of the V-PCC test models, the temporal compression of dynamic point clouds remains a highly challenging task. Due to the limitations of the current acquisition technology, the acquired point clouds can be noisy in both geometry and attribute domains, which makes it challenging to achieve accurate motion estimation. In future studies, the technologies used for 3D meshes may be exploited and adapted to provide temporal-consistent connectivity information between dynamic 3D point clouds.
Conference Paper
Full-text available
Continuous point cloud data has become very important recently as a key component in the development of autonomous driving technology, and has in fact become indispensable for some autonomous driving applications such as obstacle detection. However, such large amounts of data are very expensive to store and are difficult to share directly due to their volume. Previous studies have explored various methods of compressing point cloud data directly by converting it into 2D images or by using tree-based approaches. In this study, rather than compress point cloud data directly, we choose to compress packet data which is raw point cloud data, by converting it losslessly into range images, and then using various image/video compression algorithms to reduce the volume of the data. In this paper, four methods, which are based on MPEG, JPEG and two kinds of preprocessing approaches for each method, are evaluated for range images compression. PSNR V.S Bitrate and RMSE V.S Bitrate are used to compare and evaluate the performance of these four methods. As whether a lossy compression methods is good always depending on the application, a localization application to test point cloud data reconstruction performance is also conducted. By comparing these methods, some important conclusions can be got.
Article
Full-text available
Both ray tracing and point-based representations provide means to efficiently display very complex 3D models. Computational efficiency has been the main focus of previous work on ray tracing point-sampled surfaces. For very complex models efficient storage in the form of compression becomes necessary in order to avoid costly disk access. However, as ray tracing requires neighborhood queries, existing compression schemes cannot be applied because of their sequential nature. This paper introduces a novel acceleration structure called the quantized kd-tree, which offers both efficient traversal and storage. The gist of our new representation lies in quantizing the kd-tree splitting plane coordinates. We show that the quantized kd-tree reduces the memory footprint up to 18 times, not compromising performance. Moreover, the technique can also be employed to provide LOD (level-of-detail) to reduce aliasing problems, with little additional storage cost
Article
Full-text available
Among progressive 3D mesh compression algorithms, the kd-tree-based algorithm proposed by Gandoin and Devillers [1] is one of the state-of-the-art algorithms. Based on the observation that this geometry coder has a large amount of overhead at high kd-tree levels, we propose an octree-based geometry coder that demands a less amount of coding bits at high octree levels by applying selective cell subdivision at high tree levels, leading to a better rate-distortion performance for the low bit rate coding. Experimental results show that, compared with the kd-tree-based coder, the proposed 3D geometry coder performs better for an expanded tree of a level less than or equal to 8 but slightly worse for the full tree expansion with 12-bit quantization.
Conference Paper
Modern 3D laser scanners make it easy to collect large 3D point clouds. In this paper we present the use of conventional image based compression methods for 3D point clouds. We map the point cloud onto panorama images to encode the range, reflectance and color value for each point. An encoding method is presented to map the floating point measured ranges on to a three channel image. The image compression methods are used to compress the generated panorama images. We present the results of several lossless compression methods and the lossy JPEG on point cloud compression. Lossless compression methods are designed to retain the original data. On the other hand lossy compression methods sacrifice the details for higher compression ratio. This produces artefacts in the recovered point cloud data. We study the effects of these artefacts on encoded range data. A filtration process is presented for determination of range outliers from uncompressed point clouds.
Article
The vehicle-based mobile mapping system (MMS) is effective for capturing dense 3D data of roads, roadside objects and buildings. Since discrete points are not convenient for many application systems, a triangulated irregular network (TIN) is often generated from point-clouds. However, TIN data require two or three times larger storage than point-clouds. If TIN models can be promptly generated while loading point-clouds, it would not be necessary to store huge TIN models on a hard disk. In this paper, we propose two efficient TIN generation methods according to types of laser scanners. One is the line-by-line TIN generation method, and the other is the GPS-time based method. These methods can quickly generate TIN models based on scan lines of laser scanners. In addition, we introduce a new compression method to reduce the loading time of point-clouds. Our compression method is also based on the scan lines of laser scanners. Since points captured by a MMS tend to be positioned on nearly straight lines, their data size can be significantly reduced by coding the second order differences.
Article
Automated 3-dimensional modeling pipelines include 3D scanning, registration, data abstraction, and visualization. All steps in such a pipeline require the processing of a massive amount of 3D data, due to the ability of current 3D scanners to sample environments with a high density. The increasing sampling rates make it easy to acquire Billions of spatial data points. This paper presents algorithms and data structures for handling these data. We propose an efficient octree to store and compress 3D data without loss of precision. We demonstrate its usage for an exchange file format, fast point cloud visualization, sped-up 3D scan matching, and shape detection algorithms. We evaluate our approach using typical terrestrial laser scans.
Article
Three-dimensional models provide a volumetric representation of space which is important for a variety of robotic applications including flying robots and robots that are equipped with manipulators. In this paper, we present an open-source framework to generate volumetric 3D environment models. Our mapping approach is based on octrees and uses probabilistic occupancy estimation. It explicitly represents not only occupied space, but also free and unknown areas. Furthermore, we propose an octree map compression method that keeps the 3D models compact. Our framework is available as an open-source C++ library and has already been successfully applied in several robotics projects. We present a series of experimental results carried out with real robots and on publicly available real-world datasets. The results demonstrate that our approach is able to update the representation efficiently and models the data consistently while keeping the memory requirement at a minimum.
Conference Paper
We present a novel lossy compression approach for point cloud streams which exploits spatial and temporal redundancy within the point data. Our proposed compression framework can handle general point cloud streams of arbitrary and varying size, point order and point density. Furthermore, it allows for controlling coding complexity and coding precision. To compress the point clouds, we perform a spatial decomposition based on octree data structures. Additionally, we present a technique for comparing the octree data structures of consecutive point clouds. By encoding their structural differences, we can successively extend the point clouds at the decoder. In this way, we are able to detect and remove temporal redundancy from the point cloud data stream. Our experimental results show a strong compression performance of a ratio of 14 at 1 mm coordinate precision and up to 40 at a coordinate precision of 9 mm.