Professional Documents
Culture Documents
Abstract
Indoor environment usually has complicated structures and
contains planes which can be used as extra features for
registration. This paper proposes a method to use planar features
extracted from depth camera for indoor localization. Two
consecutive depth images are converted to point clouds and
segmented to planes. Two sets of planes are then matched
together to estimate the rotation matrix which is used as initial
guess of iterative closest point (ICP) of point-to-plane
registration. Our experimental result shows that ICP of point-toplane algorithm with the extra step of estimating the rotation
guess matrix performs faster than conventional ICP and
overcomes the drawback of ICP point-to-point algorithm in
indoor environment.
KeywordsIterative closest point (ICP), Plane segmentation,
Registration.
I.
Introduction
136
Proc. of the Third Intl. Conf. on Advances in Computing, Electronics and Electrical Technology - CEET 2015
Copyright Institute of Research Engineers and Doctors, USA .All rights reserved.
ISBN: 978-1-63248-056-9 doi: 10.15224/ 978-1-63248-056-9-85
[7, 8]. The vertical range below 1.6m from the ground is due
to the maximum height of the user sitting on a wheelchair.
Lastly, the 3D point cloud is passed through a voxel grid filter
with the leaf size of 0.05m provided by point cloud library
(PCL) [9]. A voxel is a 3D box in space and contains several
3D points, so the point cloud was divided into many square
voxel with the length of 0.05m and all points in each voxel are
represented by its centroid.
Methodology
Figure 2. Spacial filter of the system, horizotal view (left) and vertical view
(right). The valid areas are marked by green.
Figure 1. Microsoft Kinect device (left) and electronic wheelchair with Kinect
camera (right).
B.
(1)
where u and v are the row and column index of a pixel, u0, v0,
and f are the parameters configured by a depth camera.
However, depth data provided by Kinect is huge and
needs to be down-sampled and filtered before further process.
In this work, a depth image with resolution 640480 pixels is
reduced to a resolution of 320240 pixels. The spatial filter is
applied to range depth data from 0.6 4m horizontally and
below 1.6m from the floor along the vertical direction as
illustrated in Fig. 2. The depth range is chosen between 0.6
4m to ensure valid data collection and minimize acquired
errors of depth data to about 0.024m at maximum 4m of depth
137
Proc. of the Third Intl. Conf. on Advances in Computing, Electronics and Electrical Technology - CEET 2015
Copyright Institute of Research Engineers and Doctors, USA .All rights reserved.
ISBN: 978-1-63248-056-9 doi: 10.15224/ 978-1-63248-056-9-85
equation that fits to all point of the voxel. Let a set of points in
*
+
a voxel be
*
+ that is used to
formulate the plane equation
.
The
coefficients a, b, and c are found by the mean square
estimation to minimize (
)
( )
).
(2)
Step 1.
Traverse voxel grid, find a voxel Vo that has not been clustered to a
plane (non-planar voxel).
Step 2.
Search in 26 neighbor voxels Vi (i=1,2.,25,26 ) of the voxel Vo, a
plane Pi that has not been processed.
Step 3.
Calculate distance d from points of the non-planar voxel to plane
Pi. Let p (x,y,z) denote coordinates of one point in center voxel Vo, Vi
denote the i-th planar voxel in the neighborhood that belong to the plane Pi
parameterized by Ax + By + Cz + 1 = 0 .The distance from p to Pi is:
|
|
138
Figure 3. Two point clouds constructed from two consecutive depth images
shown on the same coordinate system of Kinect camera (left). The circle area
illustrates two distinguished table corners matched by a red line but they are
actually the same corner table in the real world (right).
Proc. of the Third Intl. Conf. on Advances in Computing, Electronics and Electrical Technology - CEET 2015
Copyright Institute of Research Engineers and Doctors, USA .All rights reserved.
ISBN: 978-1-63248-056-9 doi: 10.15224/ 978-1-63248-056-9-85
139
III.
Experimental Results
Figure 5. The point cloud (right) and the corresponding point cloud after plane
segmentation (left).
Proc. of the Third Intl. Conf. on Advances in Computing, Electronics and Electrical Technology - CEET 2015
Copyright Institute of Research Engineers and Doctors, USA .All rights reserved.
ISBN: 978-1-63248-056-9 doi: 10.15224/ 978-1-63248-056-9-85
Registrat
ion Fail
Mean
Iteration
3
6|4
0
n/a
4|40
3|40
22.5
4
20.4
5
IV.
Figure 6. Two point cloud after registration by ICP (left) and ICP-P2PG
(right).
Mean
Distance
Registrat
ion Fail
Mean
Iteration
SF-10
IC
P
ICPP2P
ICPP2P
G
IC
P
0.1
11
3
6|4
0
n/a
0.10
47
0.10
48
6|40
6|40
23.2
5
17.5
9
0.1
51
1
6|2
0
n/a
IC
PP2
P
0.1
65
4
4|2
0
33.
25
SF-20
IC
PP2
PG
0.1
49
8
4|2
0
33.
18
IC
P
0.2
07
2
6|1
0
n/a
IC
PP2
P
0.1
64
3
5|1
0
40.
66
IC
PP2
PG
0.1
64
2
7|1
0
12.
66
Mean
Distance
SF-10
IC
P
ICPP2P
ICPP2P
G
IC
P
0.1
19
0.11
39
0.11
22
0.2
05
IC
PP2
P
0.1
22
IC
P
0.5
70
IC
PP2
P
0.3
83
2
2|2
0
39.
14
1
5|1
0
n/a
5
5|1
0
43.
25
5
5|1
0
33.
25
Conclusion
Acknowledgment
This research is funded by Vietnam National Foundation for
Science and Technology Development (NAFOSTED) under
grant number 102.05-2013.11.
References
[1] Cox, I.J. "Blanche-an experiment in guidance and navigation of an
autonomous robot vehicle", IEEE Transactions on Robotics and Automation,
7(2), pp. 193-204, 1991.
[2] Nistr, D., Naroditsky, O., and Bergen, J., "Visual odometry, in Visual
odometry (IEEE, 2004, edn.), pp. I-652-I-659 Vol. 651
[3] Besl, P.J., and McKay, N.D., "Method for registration of 3-D shapes, in
Method for Registration of 3-D Shapes (International Society for Optics and
Photonics, 1992, edn.), pp. 586-606.
[4] Chen, Y., and Medioni, G., "Object modeling by registration of multiple
range images, in Object modeling by registration of multiple range images
(IEEE, 1991, edn.), pp. 2724-2729.
[5] https://github.com/avin2/SensorKinect/, accessed 18th December, 2013
[6] https://github.com/OpenNI/OpenNI, accessed 18th December, 2013
[7] Peasley, B., and Birchfield, S., "Real-time obstacle detection and
avoidance in the presence of specular surfaces using an active 3D sensor",
(IEEE, 2013, edn.), pp. 197-202
[8] Khoshelham, K., and Elberink, S.O., "Accuracy and resolution of kinect
depth data for indoor mapping applications", Sensors, 12 (2), pp. 1437-1454,
2012.
[9] http://pointclouds.org/, accessed 18th December,2013.
SF-20
IC
PP2
PG
0.1
23
6
5|2
0
21.
5
[10] Dub, D., Zell, A., Real-time plane extraction from depth images with the
randomized hough transform IEEE International Conference on Computer
Vision Workshops (ICCV Workshops), pp. 10841091, 2011.
6
6|2
0
n/a
IC
PP2
PG
0.4
81
140