Download as pdf or txt
Download as pdf or txt
You are on page 1of 16

Advanced Engineering Informatics 34 (2017) 1–16

Contents lists available at ScienceDirect

Advanced Engineering Informatics


journal homepage: www.elsevier.com/locate/aei

Full length article

Robust normal estimation and region growing segmentation of


infrastructure 3D point cloud models
Ali Khaloo, David Lattanzi ⇑
Department of Civil, Environmental, and Infrastructure Engineering, George Mason University, Fairfax, VA 22030, United States

a r t i c l e i n f o a b s t r a c t

Article history: Modern remote sensing technologies such as three-dimensional (3D) laser scanners and image-based 3D
Received 30 March 2017 scene reconstruction are in increasing demand for applications in civil infrastructure design, mainte-
Received in revised form 10 July 2017 nance, operation, and as-built construction verification. The complex nature of the 3D point clouds these
Accepted 11 July 2017
technologies generate, as well as the often massive scale of the 3D data, make it inefficient and time con-
Available online 2 August 2017
suming to manually analyze and manipulate point clouds, and highlights the need for automated analysis
techniques. This paper presents one such technique, a new region growing algorithm for the automated
Keywords:
segmentation of both planar and non-planar surfaces in point clouds. A core component of the algorithm
Segmentation
3D point cloud models
is a new point normal estimation method, an essential task for many point cloud processing algorithms.
Robust estimation The newly developed estimation method utilizes robust multivariate statistical outlier analysis for reli-
Outliers able normal estimation in complex 3D models, considering that these models often contain regions of
3D reconstruction varying surface roughness, a mixture of high curvature and low curvature regions, and sharp features.
Computer vision An adaptation of Mahalanobis distance, in which the mean vector and covariance matrix are derived from
Normal estimation a high-breakdown multivariate location and scale estimator called Deterministic MM-estimator (DetMM)
3D data processing is used to find and discard outlier points prior to estimating the best local tangent plane around any point
in a cloud. This approach is capable of more accurately estimating point normals located in highly curved
regions or near sharp features. Thereafter, the estimated point normals serve a region growing segmen-
tation algorithm that only requires a single input parameter, an improvement over existing methods
which typically require two control parameters. The reliability and robustness of the normal estimation
subroutine was compared against well-known normal estimation methods including the Minimum
Volume Ellipsoid (MVE) and Minimum Covariance Determinant (MCD) estimators, along with
Maximum Likelihood Sample Consensus (MLESAC). The overall region growing segmentation algorithm
was then experimentally validated on several challenging 3D point clouds of real-world infrastructure
systems. The results indicate that the developed approach performs more accurately and robustly in
comparison with conventional region growing methods, particularly in the presence of sharp features,
outliers and noise.
Ó 2017 Elsevier Ltd. All rights reserved.

1. Introduction algorithm [1], a process that converts 2D images into 3D point


clouds, is often considered as a low cost alternative to TLS systems.
Modern remote sensing technologies have enabled the genera- Accurate and efficient processing techniques of these often
tion of accurate three-dimensional (3D) geometric models that can massive 3D data is vital for extracting useful information. Point
be used to augment and improve condition assessment, progress cloud processing algorithms can be used for applications such as
tracking and documentation of civil infrastructure systems. While automatic change detection and deformation monitoring of struc-
there are many methods for creating such 3D models, terrestrial tures [2,3], or the generation of as-built 3D models in the form of
laser scanning (TLS) is the most widely used. Photogrammetric Computer-Aided Design (CAD) [4,5] or Building Information Mod-
reconstruction via the Dense Structure-from-Motion (DSfM) els (BIM) [6–8]. In many algorithms, these procedures initiate by
segmenting the point cloud data into appropriate segments and
then recognizing primitives from the segments to generate 3D
models [9].
⇑ Corresponding author.
E-mail addresses: akhaloo@gmu.edu (A. Khaloo), dlattanz@gmu.edu (D. Lattanzi).

http://dx.doi.org/10.1016/j.aei.2017.07.002
1474-0346/Ó 2017 Elsevier Ltd. All rights reserved.
2 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Segmenting point clouds of civil infrastructure systems is par- order to minimize the contribution of points located along
ticularly challenging, as these massive 3D data sets are noisy, spar- discontinuities.
sely populated, and tend to have uneven point distribution (local Recently, estimating point normals in 3D point clouds with
point density). Scanning noise is unavoidable and varies due to sharp features has been studied. Li et al. [19] implemented a statis-
laser scanning, imaging sensors, and radiometric scanning condi- tical estimator for local noise in tandem with RANdom Sample
tions. The performance of the image-based 3D reconstructions also Consensus (RANSAC) [20] to make the normal estimation proce-
vary due to camera imaging parameters, camera placement, scene dure more accurate in the presence of sharp features and planar
complexity, surface texture, environmental conditions and scene transitions. However, their method was not efficient for large 3D
size. Furthermore, the large-scale nature of most infrastructure point clouds with variations in local point density. In the work
systems, their typically complex geometry (from sharp features by Boulch and Marlet [11], a version of Randomized Hough Trans-
to highly curved regions), and the variety of textures due to varying form (RHT) was utilized for estimating point normals in point
construction materials (from smooth to very irregular), makes the clouds with sharp features. For more information on regression
segmentation of these 3D point clouds even more challenging. Seg- based methods and their performance, the reader is referred to
mentation methods that are robust in light of such variability are comparative studies and detailed analysis by Klasing et al. [21]
highly desirable. and Song et al. [22].
The focus of this paper is on the development and validation of Within civil engineering applications, it is possible to improve
a point cloud segmentation algorithm that is designed with these the performance of many normal vector-based 3D model analysis
considerations in mind. The algorithm uses a region growing techniques such as automated crack detection [23], deformation
approach to isolate and segment both planar and non-planar sur- measurement [24] and 3D as-built modeling [25,26], through a
faces in large scale 3D point clouds. As a first step in this larger more accurate estimation of point normals.
algorithm, point normals for each point in a cloud are estimated.
A new and more robust outlier detection process has been 2.2. Prior work in point cloud segmentation
designed to improve these estimations.
Segmentation algorithms can be roughly classified into three
categories: Model fitting [27], Clustering [28] and Region growing
2. Related work [29] methods.
Model fitting-based methods attempt to find sets of points in
The state of research in as-built 3D modeling of infrastructure the scene that fit 3D primitives. These methods generally use dif-
systems and semi-automated segmentation routines were covered ferent adaptations of well-known algorithms such as Hough Trans-
in comprehensive review studies by Pătrăucean et al. [8] and Son form (HT) and RANSAC [30]. In most cases these techniques are
et al. [10], respectively. Previous work in this domain can be sepa- able to identify surfaces in cluttered scenes. However, they lever-
rated into studies on normal estimation and/or point cloud age shape primitives and are thus context specific. These methods
segmentation. have been used to segment point clouds of civil infrastructure sys-
tems [31,32], but tend to perform poorly in the face of large data
sets or complex geometries [30] and were not considered further
2.1. Prior work in normal estimation in this study.
Clustering-based methods aim to segment sets of points based
Reliable estimation of the normal vector at each point in a 3D on clustering local feature attributes of 3D point cloud data [33].
point cloud has become a fundamental step in point cloud data These techniques are computationally expensive for multi-
processing. Numerous algorithms rely on accurate normal estima- dimensional features in large datasets and their performances are
tion, such as point-based rendering, surface reconstruction, 3D heavily impacted in the presence of a high proportion of outliers
piecewise-planar reconstruction, and 3D point cloud segmentation [28]. To date these approaches do not perform well on points
[11]. Therefore, methods that provide a good trade-off between placed on discontinuities (e.g. sharp features and planar transi-
computational efficiency and accuracy are essential. Normal esti- tions), since the extracted feature vectors of these points often dif-
mation in 3D point clouds can be classified into Voronoi/Delaunay fer from other points located on adjacent planes.
and Regression methods. The comparative study by Dey et al. [12] Region growing-based methods [29] use local features
showed that Voronoi methods are unreliable in noisy point clouds extracted from a neighborhood around each point to aggregate
or in clouds with variations in local density variation, making this nearby points with similar properties and segment a region of a
class of methods unsuitable for civil infrastructure applications. point cloud. Rabbani et al. [34] proposed a region growing tech-
Regression based normal estimation was pioneered in the work nique based on smoothness constraints to extract smoothly con-
by Hoppe et al. [13], as a module in their surface reconstruction nected areas in unstructured 3D point clouds. This method uses
algorithm. Their method estimated the normal at a point p by fit- surface normal vectors and approximated local curvature values
ting a local plane under the assumption of planarity in the region of (which depends on normal estimation and orientation) as mea-
the k-nearest neighbors of a given point, using Principal Compo- sures of local geometry to determine each segment. Since the nor-
nent Analysis (PCA) [14]. Due to the sensitivity of PCA to outliers mal estimation in this approach was calculated by using the PCA-
and the implicit smoothness assumption, this method is sensitive based method [13], the algorithm is susceptible to noise and sharp
to outliers and noise, and cannot handle sharp features. To avoid features within the point cloud model. In the work by Son et al.
strong assumptions for a local 3D neighborhood and reduce the [35], the aforementioned method was implemented to automati-
sensitivity of PCA, both Mitra et al. [15] and Wang et al. [16] sug- cally extract 3D points corresponding to pipelines in an industrial
gested an adaptive neighborhood size based on an iterative scheme plant. Walsh et al. [36] also implemented this algorithm to seg-
to find the optimal k closest 3D points. Pauly et al. [17] assigned ment point clouds of small structural elements. In both of these
Gaussian weights to distances between point p and its k-nearest studies, the targeted 3D scanned data was relatively ‘‘noise-free”
neighbors. To make the normal estimation more consistent near and without any major challenges such as missing data due to
geometric singularities, such as corners and edges, Castillo et al. occlusion or significant variability in surface roughness and curva-
[18] formulated PCA as a constrained nonlinear least squares prob- ture. Inspired by [34], Dimitrov and Golparvar-Fard [9] introduced
lem (NLSQ) to assign appropriate weights to neighboring points in a region growing segmentation for building point clouds that
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 3

leverages multi-scale feature detection, such as roughness and cur- ality of the data. For dense point clouds of infrastructure systems,
vature, for each 3D point. which often contain hundreds of millions of points, computing the
k-NN of every point is computationally prohibitive. In order to
make this process more efficient, the k-d trees space partitioning
2.3. Contributions of this work
method is applied to the k-NN search [39], to reduce the search
complexity to Oðlog nÞ.
As shown by Yoon et al. [37], inaccurate estimation of point
normals is the most likely source of errors in segmentation,
emphasizing the need for advancements in normal estimation 3.1. Normal and curvature estimation
algorithms, in particular for civil infrastructure system models
with a high number of sharp planar transitions. Current region After determining the k-NN for a point p, the approximate point
growing methods also require multiple user defined control normal is then estimated. The salient local 3D features, the point
parameters, creating an algorithm sensitive to user inputs and sus- normal vector and curvature, are first estimated for each point p.
ceptible to variations in surface roughness and cloud density. Over- The conventional approach to normal estimation is to use Principal
all, there is a need to develop accurate and efficient normal Component Analysis (PCA) [13], with the point normal estimated
estimation and segmentation algorithms capable of handling large as the eigenvector associated with the smallest eigenvalue of the
scale and noisy point clouds of civil infrastructure systems, which symmetric positive semi-definite matrix:
often contain both highly planar and curvilinear objects.
1X k
1X k
This paper presents a new region growing method that C 33 ¼ ÞT ðpi  p
ðp  p Þ; ¼
p p ð1Þ
k i¼1 i k i¼1 i
addresses the challenges of sharp features (planar transitions),
gaps due to occlusion and missing data, and outliers/noise. An effi-  is the local data centroid
where C is the covariance matrix and p
cient and robust statistical outlier detection process is imple-
(arithmetic mean) within the neighborhood N pi of point p. By per-
mented to more accurately estimate normal vectors near sharp
forming Singular Value Decomposition (SVD) [40] on the covariance
features, and in the presence of varying levels of noise and sam-
matrix, it is possible to compute the eigenvectors V (v 2 ; v 1 and v 0 )
pling density. These improved estimates of point normals are then
and their corresponding eigenvalues k (k2  k1  k0 ). Within this
used in a region growing segmentation that requires only a single
approach v 0 approximates the point’s p normal, while v 2 and v 1
input parameter, minimizing parameter sensitivity. This is accom-
span the tangent plane at p.
plished through the embedment of a locally adaptive criteria
In addition, k0 quantitatively describes the variation along the
within the region growth step, providing an intrinsic ability to han-
point normal, providing an estimate of the local noise level (a.k.a
dle large variations in surface roughness without the need for a
surface roughness) as a description of how much the k points devi-
control parameter.
ate from the fitted tangent plane. The smaller the value of k0 the
The performance of the developed normal estimation algorithm
better the quality of the fitted plane. Curvature (normalized sur-
was first compared against a number of well-established alterna-
face variation) can be defined as [17]:
tive processes such as the PCA-based approach [13]. The overall
segmentation algorithm was then compared with the existing k0
rðpÞ ¼ k2 P k1 P k0 ð2Þ
region growing approach implemented in the Point Cloud Library k0 þ k1 þ k2
(PCL) [38]. The results of these comparative studies illustrated that
the new segmentation approach has a higher robustness in the if rðpÞ ¼ 0, then all the points lie in the perfect plane.
presence of noise, sharp features, and high curvature regions. The PCA approach is relatively accurate when the k-NN of a
point p is located on a smooth surface, but tends to smear across
singularities such as sharp features or noise due to the contribution
3. Methodology to the PCA covariance matrix, C, of neighboring points across the
singularity (Fig. 2). This limitation then results in inaccurate seg-
The overall segmentation algorithm (Fig. 1) initiates by first mentations at the region growing stage. The challenges of sharp
identifying the k-nearest neighbors (k-NN) for every point in the features and scanning noise are particularly significant for point
dataset. The linear search solution for k-NN has a running time clouds of civil infrastructure systems, which tend to exhibit many
of OðndÞ, where n is the number of points and d is the dimension- of these issues.

Fig. 1. The automated robust segmentation process.


4 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 3. Comparison of estimated normals on the CUBE dataset.

Fig. 2. Normal estimation challenges.

th
xi ¼ ðxi1 ; . . . ; xim ÞT is the i observation within that neighborhood.
However given an accurate normal estimation, sharp features The observation fxi g is thus a potential outlier under investigation,
will be preserved and segmented surfaces will correctly represent P
where flx g ¼ 1n ni¼1 xi is the arithmetic mean of the sample obser-
the physical object geometry. The method developed here accom- P
plishes this through the use of statistical outlier detection to detect vations, and R ¼ ni¼1 ðxi   xÞðxi  xÞT =ðn  1Þ is the sample covari-
singularities and outliers prior to choosing an appropriate set of ance matrix.
inlier neighboring points for PCA. A new multivariate statistical This approach works well if only a single outlier is present, but
approach called Deterministic MM-estimator (DetMM) is used to suffers from a masking effect otherwise, because one far-away out-
determine inlier points within each N pi . The DetMM approach is lier can reduce D2i for other outliers [42]. This effect can be mea-
robust in the presence of noise and geometric complexities, includ- sured by a metric known as the breakdown point. The breakdown
ing sharp features (Fig. 3). point is the smallest fraction of outliers that may cause an arbitrar-
ily large effect on the estimation of the regression coefficients. For
3.2. DetMM-based Normal Estimation (DMNE) example, the breakdown point for the Mahalanobis distance is only
2/n, which makes it extremely sensitive to outliers.
The classic distance-based discordancy measure for indicating The problem of masking effect can be resolved by using robust
outliers is the squared Mahalanobis distance [41]. The (squared) estimates of mean and covariance matrix that are less affected by
Mahalanobis distance is similar to the Euclidean distance, except outliers. In the DMNE approach (Fig. 4), a variant of the squared
that it normalizes the data on the basis of the inter-attribute Mahalanobis distance is implemented to increase the breakdown
correlations. point of the distance estimate. Here, the deterministic multivariate
MM-estimator (DetMM) is used to compute the mean vector ( l b)
D2i ¼ ðfxi g  flx gÞT ½R1 ðfxi g  flx gÞ i ¼ 1; . . . ; n ð3Þ b
and covariance matrix ( R ):
T 1
A multivariate matrix ½X ¼ ðx1 ; . . . ; xn Þ describes an m- D2i ¼ ðfpi g  f l
T b  ðfp g  f l
b gÞ ½ R b gÞ 6 v2m;0:975 i ¼ 1; . . . ; n ð4Þ
i
dimensional local point neighborhood of size n, where
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 5

Fig. 4. The DetMM-based normal estimation (DMNE) process.

1=2
As suggested by Rousseeuw and Leroy [42], hard-rejection 1X n
ððx  mÞT C1 ðxi  mÞÞ
weights are utilized by omitting any point whose estimated q1 ð i Þ ð7Þ
n i¼1 r~
squared Mahalanobis distance is larger than the 97.50 percentile
of the chi-square (v2 ) distribution with m = 3 degrees of freedom Over ðm; CÞ where m 2 R3 and C is a PDS matrix with jCj ¼ 1.
(v23;0:975 ¼ 3:075). The minimization is performed through iterative least-squares,
yielding ðl ^ Þ. The resulting DetMM estimator of mean and
^; C
covariance matrix are l ^ ¼r
^ and R ^ , respectively.
~ 2C
3.2.1. Deterministic MM-estimator (DetMM)
Similar to the MM-estimators of regression introduced by Yohai
[43], MM-estimators of multivariate location and shape [44] are 3.2.2. Robust normal estimation
based on two loss functions (q0 and q1 ). The following regularity Upon finding inliers for the local k sized neighborhood N pi of a
conditions are required for a loss function q: point pi , PCA then is applied to the inlier points in order to deter-
mine the best fit tangent plane for the cleaned neighborhood, as
(i) q is real, symmetric, twice continuously differentiable, and described previously (Fig. 4).
qð0Þ ¼ 0: Since the orientation of the estimated normals are ambiguous, a
(ii) q is strictly increasing on ½0; c for some c > 0 on ½c; 1½ for globally consistent orientation must be computed in a separate
some finite constant c. step. This can be accomplished through the use of a minimum
spanning tree (MST) approach to reorient raw normals using the
The loss function q can be defined as: term 1  jhni ; nj ij as the uncertainty measure for this
propagation-based approach, where ni;j are the two normals of
( adjacent points [13]. The key idea in this approach is to consider
x2 4 6
 2cx 2 þ 6cx 4 for jxj 6 c
qðxÞ ¼ 2
c2
ð5Þ that two data points pi and pj belonging to a smooth surface and
6
for jxj > c geometrically close, need to have their normals consistently ori-
ented (i.e. ni  nj 1). Although this MST-based approach works
Multivariate MM-estimators of the location, shape, and covari-
well for smooth and densely sampled surfaces with low curvature,
ance of the point neighborhood ½X ¼ ðx1 ; . . . ; xn ÞT  Rp it can fail at sharp edges and creases. In order to consistently exe-
(p-dimensional) can be robustly estimated by initially utilizing cute the re-orientation of the calculated normals in large noisy 3D
S-estimators [45]. An S-estimator of location and scatter is defined point clouds, the adaptation presented by Schertler et al. [47] was
as the couple ðl ~ Þ which minimizes jSj under the condition:
~; R implemented due to its performance in the presence of noise and
sharp features.
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
1X n Fig. 3 illustrates the benefits of using outlier invariant normal
q ðxi  mÞT S1 ðxi  mÞ ¼ b ð6Þ estimation in comparison with the conventional PCA approach,
n i¼1 0
for a synthetically generated 3D point cloud (a perfect cube) with
Among all ðm; SÞ where m 2 R3 and S is a 3  3 positive definite sharp features. Unlike the conventional PCA approach, which does
symmetric (PDS) matrix with b ¼ 1n ½npþ1 . Recently, Hubert et al. not reject outliers, the DetMM normal estimation method is able to
2
reliably estimate point normals near the planar transitions of the
[46] proposed a deterministic S-estimator (DetS) that uses a
cube.
resampling strategy to improve the efficiency of the algorithm,
Due to the more robust estimation of point normals using
and their approach is used here.
DMNE, the values of the surface variation, rðpÞ, are more consis-
Deterministic MM-estimators (DetMM) [46] can then be
tent even for non-planar smooth surfaces (Fig. 5), when compared
derived by utilizing the DetS approach as the initial estimator for
with conventional PCA. Fig. 6 illustrates the benefits of the pre-
a multivariate MM-estimator, defined in two steps:
sented method. The boxes enclose the middle half of the results,
  i.e. the length of a box is the interquartile range with the ends of
(i) Let ðl; RÞ be a DetS estimator with loss function q0 . Denote the box at the lower/1st and upper/3rd quartiles. The line across
~ j1=ð2pÞ
r~ ¼ jR each box is the position of the median, and the ‘+’ markers repre-
(ii) The DetMM-estimator for location and shape ðl ^Þ
^; C sent outlying results. The box length shows the variability of the
minimizes: data.
6 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 5. Comparison of estimated normalized surface variation (curvature) for a point cloud of a pipe.

3.3. Alternative methods used for comparison

Alternative estimators of multivariate location and scatter


include: S-estimators [45], s-estimator [48], Minimum Volume
Ellipsoid (MVE) [49] and Minimum Covariance Determinant
(MCD) [42] estimator. They are all highly robust estimators with
breakdown points up to 50%, but they differ in terms of efficiency,
bias and outlier resistance. For detailed comparison of these
approaches, the reader is referred to [50]. In this work, the follow-
ing alternative methods were assessed relative to the utilized
DetMM approach.

3.3.1. Minimum Volume Ellipsoid (MVE) and Minimum Covariance


Determinant (MCD)
One of the earliest affine equivariant estimators to achieve a
breakdown point of approximately 50% is the Minimum Volume
Ellipsoid (MVE) estimator. The MVE estimator [49] of location
(mean) is defined as the center of the ellipsoid with minimal vol-
ume containing h out of m points (any n dimensional) with
½m=2 6 h 6 m. The corresponding scatter estimator (covariance
matrix) is given by the formula for this ellipsoid, multiplied by a
qffiffiffiffiffiffiffiffi
consistency factor c. Where c ¼ v2n;a with a ¼ h=m. According to
Fig. 6. Boxplot of curvature values for the pipe model. its definition, MVE estimates the mean and covariance of the h
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 7

most concentrated points in the dataset. This estimator has maxi- point is needed for the growth of the next segment. In this study,
mal breakdown value when h ¼ ½ðm þ n þ 1Þ=2 but it is not the point pi with the minimum curvature value, rðpÞ, was consid-
asymptotically normal and more difficult to compute than the ered as the initial seed to start the region growing process from a
MCD estimator [42]. part in the scene where the surface is smoother and, hence, the
An alternative to the MVE estimator that also has a breakdown surface variation is lower.
point of approximately 50% is the minimum covariance determi-
nant (MCD) estimator [42]. Rather than search for the subset of 3.4.1. Spatial connectivity analysis in segment growth
half the data that has the smallest volume, MCD searches for the While it has been used effectively in previous work, the
half h > m/2 whose classical covariance matrix has the lowest pos- smoothness constraint based region growing method of Rabbani
sible determinant. Within the MCD approach, l is the mean of the et al. [34] suffers from several shortcomings. Firstly, due to the
h observations, ½ðm þ n þ 1Þ=2 6 h 6 m, for which the determinant PCA based normal estimation strategy implemented within, it is
of the sample covariance matrix is minimal and the scatter matrix inherently susceptible to noisy outliers. Secondly, this algorithm
R is the corresponding covariance matrix multiplied by a consis- requires two parameters/thresholds (for both curvature rðpÞ and
tency factor in the case of multivariate normality. Generally, the normal vector angular distance hij values). As suggested in
MCD estimates of location and scatter are affine equivariant which [36,34,35], choosing the curvature threshold as the 98th percentile
means that they are also invariant under affine transformation of the sorted normalized surface variation makes the segmentation
behavior. It should be noticed that the MCD estimator can only results data dependent. Fig. 7 illustrates the segmentation of a 3D
be computed when h > n, otherwise the covariance matrix of any table model using this method [34]. Smooth areas are broken by
h-subset will be singular. Since h P ½ðm þ 2Þ=2, this condition is using solely the angular distance criterion between two adjacent
certainly satisfied when m  2n. points. On the surface of the object due to the sensitive nature of
Although MCD is easier to compute in comparison with MVE, it the DSfM process to environmental factors (e.g., humidity, temper-
is still impractical for very large data sets since it requires the eval- ature and lighting) and lack of adequate lighting while collecting
 
m image data under the surface, higher noise levels were achieved
uation of subset of size h. In order to make this approach
h at these locations which later resulted in over-segmentation and
scalable and suitable for large 3D point cloud datasets, an efficient failing of the curvature threshold.
version of MCD with fast resampling called FAST-MCD algorithm Inspired by [54], and in order to overcome this difficulty, the
[51] was implemented in this study. segmentation algorithm developed herein uses a single user-
defined parameter and two automatically calculated locally adap-
3.3.2. Maximum Likelihood SAmple Consensus (MLESAC) tive thresholds to account for variations in the local geometrical
RANSAC is a robust approach to classify data into inliers and features of each point.
th
outliers which is implemented in many applications for extracting Three distance criteria are defined between the i seed point pi
shapes and estimating the model parameters from data that con- and the k-sized neighborhood, N pi , and used to add points to the
tain noise/outliers [27,19]. Compared to RANSAC, MLESAC algo- growing region. The test criteria are: (i) Coplanar Distance (CD),
rithm adopts the same sampling strategy but chooses the (ii) Euclidean Distance (ED) and (iii) Angular Distance (h).
solution by minimizing the probabilistic loss rather than the num- Point-to-plane coplanar distance (CD) measures the distance for
ber of inliers. Minimizing probabilistic loss renders the model ver- every neighboring point N pi of the pi to its best-fitted tangent local
ification threshold value invariant to the cardinality of the models plane using:
dataset. MLESAC is reported to be of higher accuracy and robust-
ness than RANSAC [52], which makes it more suitable for local  Þ T :n
CDðpi Þ ¼ ðpi  p ^ ð9Þ
plane fitting. In this approach, first a minimal subset, e.g. three
where p  and n b are the local data centroid (mean) and the unit nor-
points for a plane, is randomly selected and the required plane
mal of the locally fitted plane, respectively.
parameters are then computed based on this subset. Next, MLESAC
In order to only add salient points during region growth, it is
evaluates the fitness of the hypothesized parameters using a prob-
important to define a robust threshold to identify outliers. As pre-
abilistic model for the errors arising from noise contaminated data.
viously used for noise/outlier detection in point clouds [36,9], it is
As discussed in [53,32], MLESAC measures the likelihood of the
possible to utilize mean (l) and standard deviation (r) in the form
generated model hypothesis by representing the error distribution
of l þ 2r to distinguish inliers from outliers. Although mean is the
as a mixture model. The suggested probability distribution func-
classic estimation of location, it is very sensitive to even one aber-
tion (pdf) for inlier error is unbiased Gaussian distribution and
rant value out of the n data points which makes its breakdown
the pdf of outlier error is uniform distribution. Thus the probability
value equal to 1=n 0% for large n. In addition, standard deviation
of error described as follows: qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Pn ffi
      r¼ 2
i¼1 ðxi  xÞ =ðn  1Þ also has a breakdown point of 0% since
1 e2 1
PrðejModelÞ ¼ c pffiffiffiffiffiffiffiffiffiffiffi2ffi exp  2 þ ð1  cÞ ð8Þ a single outlier can make r arbitrarily large.
2pr 2r m To address this, median and Q n estimator [55] are implemented
where r is the standard deviation of Gaussian noise, e is the inlier instead of mean and standard deviation to make the threshold for
error, c is the prior probability of being an inlier and m is the size the coplanar distance (CDth ) robust in the presence of noise. In con-
of the available error space. When the PrðejModelÞ is larger than a trast with mean, median has a breakdown value of about 50%,
predefined threshold, then the model will be reestimated using only meaning that it can resists up to 50% of outliers. Q n estimator is
inliers and MLESAC terminates. a robust measure of scale which is defined as the first quartile of
all pairwise differences between two data points:
3.4. Region growing segmentation Q n ¼ 2:2219fjxi  xj j; i < jgðkÞ ð10Þ
   
Once the point normals and curvatures are estimated, the h n
with k ¼ 4 and h ¼ bn=2c þ 1. The constant 2.2219 is
region growing segmentation starts by finding a seed point pi that 2 2
is then grown by incrementally adding new points to it using the a correction factor that makes the estimator unbiased. Q n estimator
region growing test criteria. When growth is complete, a new seed has a breakdown value of 50% and complexity of Oðn log nÞ.
8 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 7. Estimated curvature values and segmentation of the Table model using Rabbani et al. [34] method with 98th percentile curvature values.

As discussed previously in the work by Li et al. [19], noise Consequently, pj is considered to be spatially connected to pi if
within 3D point cloud data obeys Gaussian distribution. In compar- in comparison with the majority of the points in N pi they are rela-
ison with other robust scale estimators such as median absolute tively close:
deviation (MAD) and Sn (which has Gaussian efficiency close to
37% and 58%, respectively), Q n estimator has a high efficiency of EDij < EDth ¼ medianfEDij g ð13Þ
82% at Gaussian distributions [55] which makes it highly suitable
th
where EDth is the ED threshold and fEDij g is the set of all EDij s.
for point cloud analysis. Hence, the i point is considered a copla- th
Finally, the angular distance is defined between the i seed
nar surface point if:
point pi and one of its neighbors pj as:
CDi < CDth ¼ medianfCDðNpi Þg þ 2  Q n fCDðNpi Þg ð11Þ
^ Ti :n
hij ¼ arccos jn ^j j ð14Þ
where fCDðNpi Þg is the set of CDs for all points in Npi .
The data points that have the closest distance to each other are where n ^ i and n
^ j are the unit normals for points pi and pj , respec-
most likely to be within the same segment on the 3D point cloud. tively. As the direction of normal vectors has a 180 ambiguity, here
Point-to-point Euclidean Distance (ED) computes the distance the absolute value of the dot product was used. This distance mea-
between a pair of points (the seed point pi and one of its neighbors sure (h) plays a pivotal role in segmenting 3D point clouds at inter-
pj in N pi ) as: section edges were two surface patches meet. Two spatially close
EDij ¼ kpi  pj k ð12Þ points are co-plane if h is less than the user defined threshold hth .
Due to the robustness of the estimated normal vectors near sharp
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 9

features and surface discontinues through using the DetMM based order to quantify normal estimation sensitivity and (iv) computa-
method (Fig. 3), the presented region growing segmentation is more tional time.
efficient in recovering sharp features in the presence of bilinear sur- The efficiency and robustness of the DMNE algorithm was
faces. The hth threshold is the only user-dependent parameter (as tested on point cloud datasets with various geometric features,
CDth and EDth are locally adaptive and automatically derived) noise levels and curvature values. The S1 model (Fig. 9) contains
within the presented algorithm, and can be intuitively selected highly curved pipes and support structures from a laser scanned
based on the expected geometric complexity of a scene with respect industrial facility scene with varying local point density, and con-
to the presence of sharp planar transitions. sists of 195,057 points. This model is from a publicly available
Fig. 8 shows an example segmentation result and curvature val- dataset by RAAMAC lab at the University of Illinois at Urbana
ues based on the presented method. In contrast with the result Champaign. The other dataset ‘‘Table” (154,100 points) was gener-
achieved using the method in [34] (Fig. 7), a much more accurate ated by using a hierarchical adaptation of the DSfM method [56],
segmentation with improved accuracy in regions with a relatively and consists of both planar and non-planar complex objects with
higher noise level is apparent. This illustrates the robustness of the variations in surface roughness (Fig. 9).
algorithm in segmenting 3D datasets consisting of planar and non- The average local point cloud density values for each of these
planar complex object surfaces while preserving the overall geom- datasets are 50 points/m3 and 28 points/m3 , respectively. It is
etry of the reconstructed physical object. important to note that the k sized neighborhood for each point p
The complete region growing segmentation (Fig. 1) algorithm is should be large enough to negate noise while still small enough
provided as pseudocode for repeatability (Algorithm 1). to accurately estimate local properties. While greater neighbor-
hoods certainly help suppress noise in the sense that they smooth
Algorithm 1. The complete segmentation method with the out the estimated surface, they do not necessarily yield a better
embedded DMNE process surface normal estimate. In this paper k = 50 was used to fit a plane
to a local neighborhood at each 3D point.
Tables 1–3 shows the results of normal estimation and outlier
detection using a variety of estimators. The comparative analysis
demonstrates that, by finding and discarding outliers within the
local neighborhood of each point, it is possible to more accurately
estimate normal vectors for noisy point clouds by utilizing the
DMNE method. A disadvantage of utilizing MCD and MVE as mul-
tivariate robust estimators to detect and discard outlier points
within each local neighborhood, as illustrated in Fig. 10, is that
they tend to declare a larger number of outliers when the sample
size is small. This is because they make substantial use of the
asymptotic v2 approximation to the distribution of robust dis-
tances which may result in inaccurate results even in moderately
large datasets [42].
In addition, in this comparative study the best tangent local
plane was fitted for every point before and after contaminating
datasets by adding synthetic Gaussian noise with zero mean and
0.2 variance. The approach presented herein performs better (i.e.
smaller bias angles) than other methods in the presence of this
noise contamination (Table 1). Results in Table 3 show that the
DMNE algorithm is significantly faster than other methods in find-
ing the best-fit plane within each points neighborhood. This com-
putation can be performed in parallel using multiple threads to
take advantage of multiple CPU cores, although this was not done
in this work.

4.2. Evaluating segmentation accuracy

To determine the efficiency and robustness of the presented


4. Performance analysis point cloud segmentation algorithm, it was tested against the
smoothness constraint region growing method proposed by [34]
Both the DetMM-based normal estimation process (DMNE) and which has been implemented in C++, and published in the PCL
the overall segmentation algorithm were evaluated. In this study [38]. To validate the performance of the point cloud segmentation
the implementation of the entire code was done in MATLABÒ, run- algorithms, a set of ground truth 3D data were manually generated
ning on a MicrosoftÓWindows computer with a 3.4 GHz Intel i7- by human experts, who were familiar with the surface segmenta-
3770 processor and 32 GB RAM. tion process, to serve as a reference dataset. These reference seg-
ments were extracted based on experts’ judgment of surface
4.1. DMNE evaluation continuity and not object continuity.

In order to assess the performance of the DMNE process, the fol- 4.2.1. Performance metrics
lowing metrics are used: (i) Variation along the point p normal, There are a number of available methods in the literature to
defined as the least eigenvalue k0 , (ii) curvature or normalized sur- evaluate the agreement between two sets of segmented data.
face variation rðpÞ, (iii) bias angle h ¼ arccos jn^ Ti :n
^ j j, between the Herein, the metrics precision, recall and F-score were utilized to
best fitted local planes with and without added noise/outliers in measure the overall accuracy and effectiveness of a segmentation.
10 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 8. Estimated curvature values and segmentation of the Table model using the presented method with hth ¼ 4 .

These metrics are based on the following values: (i) True Positives Finally, the F-score is the harmonic mean of precision and recall
(TP) represents proper segmentation that matches with a manually values, summarized in a single factor:
generated ground truth benchmark by human experts; (ii) False
Positives (FP) denotes over-segmentation within point cloud mod- Precision  Recall
els that results in segmenting one ‘reference’ segment into several F-score ¼ 2  ð17Þ
Precision þ Recall
parts; (iii) False Negatives (FN) happen when two or more separate
segments are wrongly grouped together to under-segment the 3D Due to the sensitivity of the precision and recall metrics to the
model. existence of spurious elements and ‘reference’ data that are not
Precision represents the capability of the algorithm in properly recognized by the algorithm, respectively, the F-score represents
segmenting 3D point clouds, as given by: the best metric to evaluate the robustness and effectiveness of
the segmentation process.
jTPj
Precision ¼  100 ð15Þ Since manually comparing extracted segments of point clouds
jTPj þ jFPj against a ground truth dataset of large infrastructure 3D models
The recall metric allows to measure the ability of the 3D seg- is not a straightforward task, the automated object-based evalua-
mentation algorithms to correctly retrieve ‘reference’ segments tion method proposed by Awrangjeb and Fraser [57] was imple-
according to the ground truth data: mented in this study. This method automates the process of
making one-to-one correspondences between extracted and refer-
jTPj ence segments of point clouds using the maximum overlaps. First,
Recall ¼  100 ð16Þ
jTPj þ jFNj it establishes a pair of correspondences between extracted seg-
ments, S ¼ fS1 ; S2 ; . . . ; Si g, and ground truth segments,
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 11

G ¼ fG1 ; G2 ; . . . ; Gj g. In this step, ðSi ; Gj Þ is considered a proper one-


to-one correspondence if each entity Si provides the maximum
overlap with Gj among all extracted parts and vice versa.

4.2.2. Experimental validations and discussions


A series of segmentation experiments were designed and run on
several complex 3D point clouds with different geometric charac-
teristics to evaluate the performance of the introduced algorithm
in comparison with the PCL implementation of [34], referred to
as PCL-RG (hth ; rth ). hth and rth are the angle threshold (in degrees)
and curvature threshold, respectively. Here, the rth were calculated
automatically using a specified percentile of the sorted curvature
values as suggested in [34–36].
In addition to the S1 and Table point clouds used for DMNE
evaluation, two corridor scenes (models T1 and T2), each contain-
ing densely packed plumbing equipment and featuring a signifi-
cant amount of occlusion and clutter, were used. T1 (554,150
points, density of 1230 points/m3 ) and T2 (1,360,945 points, den-
sity of 5768 points/m3 ) models, along with the S1 model were
selected to evaluate the segmentation algorithms. Fig. 11 illus-
trates the segmentation results on the T2 model. Due to occlusions
and the scan registration process, a large range of surface rough-
ness is apparent within this model which results in variability in
the normalized noise level. As shown in Fig. 11, there is no clear
set of input parameters for the PCL-RG method that can be deemed
best. In case of scenes containing variable levels of curvature,
where multiple objects are in close proximity, finding the proper
set of parameters will inevitably be a challenge. According to
Fig. 11, by setting the rth as the 95th percentile of the curvature
values, a noticeable amount of over-segmentation occurred in loca-
tions with excessive noise. Increasing the aforementioned user-
defined parameter to 98th percentile, resulted in an inappropriate
under-segmentation by falsely including points from different sur-
faces as one. The results from the presented algorithm reveal that
by incorporating DetMM-based normals and locally adaptive
thresholds using the spatial connectivity analysis, it was possible
to properly extract surfaces with various curvature values and even

Table 2
Comparison of accuracy in tangent plane fitting based on different methods.

Data set Methods Mean curvature rðpÞ Mean noise level k0 (mm)
S1 PCA 0.0128 3.7
MVE 0.0022 0.8
MCD 0.0027 1.1
MLESAC 0.0041 1.2
DetMM 0.0015 0.6

Fig. 9. 3D point cloud datasets. Table PCA 0.0615 21.1


MVE 0.0195 7.01
MCD 0.0251 9.3
MLESAC 0.2915 10.14
DetMM 0.0165 4.6

Table 1 Table 3
Bias angles h (in degrees) for different methods. Comparison of runtime in tangent plane fitting based on different methods.

Data set Methods Mean Median Data set Number of points Methods Execution time (minutes)
S1 PCA 21.38 12.88 S1 198,217 PCA 0.11
MVE 11.62 9.12 MVE 3187.21
MCD 12.71 12.26 MCD 1487.67
MLESAC 14.13 12.18 MLESAC 294.65
DetMM 7.28 5.61 DetMM 56.77
Table PCA 18.74 10.81 Table 154,100 PCA 0.09
MVE 8.42 7.35 MVE 2645.43
MCD 11.18 7.73 MCD 1207.25
MLESAC 15.83 9.45 MLESAC 274.47
DetMM 3.14 2.18 DetMM 40.12
12 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 10. Index plot of robust Mahalanobis distances calculated using (a) MVE, (b) MCD and (c) DetMM estimators.

close parallel and adjacent surfaces of different sizes using a single Table 4 presents the quantitative assessment of different seg-
parameter segmentation. mentation algorithms. Overall, the introduced method yielded con-
Figs. 12 and 13 illustrate that the new method is capable of sistently better results for precision (i.e. correctness), recall (i.e.
cleanly separating objects in close proximity in both T1 and S1 completeness) and F1-score (i.e. overall accuracy) in all 3D data-
models. The higher accuracy and robustness of the developed sets. Minimizing user defined parameters and considering the geo-
method improves the segmentation quality in smooth regions metric properties of the 3D point cloud data resulted in significant
while preserving sharp features around corners and close to other improvement in adequately segmenting multiple objects in close
objects. The over-smoothing of point normals near or on sharp cor- proximity. The overall lower F1-score values (Table 4) for the T1
ners with the PCA-based approach limits the performance of the and T2 models is due to the relatively higher rate of occlusions
conventional region growing method in recovering points on these and missing data in these models.
locations. The segmentation results also show the capability and robust-
Since the region growing method implemented in the PCL was ness of the presented algorithm in dealing with a large range of
utilized in this comparative study, the suggested point cloud local point density and surface variation, a condition prevalent
posted on [38] for benchmarking purposes was used to test the for many point clouds of large infrastructure systems. Notably, as
new method. This synthetic 3D point cloud (Fig. 14) has 154,100 shown in Figs. 8 and 11, the PCL-RG algorithm displayed poor per-
points and a local point density of 28 points/m3 and contains both formance in over- or under-segmentation of point clouds with
sharp and smooth transitions between mostly planar surfaces. As varying surface roughness due to considering curvature (rth ) as
illustrated in Fig. 14, PCL-RG had the tendency of splitting a one of the user-defined values.
smooth patch into many smaller segments where even small gaps
appeared (e.g. close-up views in Fig. 14) in the final segmented 5. Conclusions
model. In comparison, the developed algorithm successfully
extracted closely connected planes with a variety of angular dis- This paper presents a new and statistically robust method for
tances between them due to the less noticeable gaps between seg- reliable normal vector estimation in noisy 3D point clouds with
mented surfaces. non-uniform point distribution and high variance in the surface
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 13

Fig. 11. Segmentation of a corridor scene (model T2) with various surface roughness.

Fig. 12. Segmentation of a corridor scene (model T1) with various surface roughness.

roughness and geometric complexity, referred to as DMNE. DMNE points positioned on the same smooth surface in normal
incorporates a deterministic and efficient variant of a high break- computation.
down MM-type multivariate location and scatter estimator to Experiments based on several complex 3D point cloud datasets
improve the robustness of the Mahalanobis distance-based tech- generated using laser scanning and photogrammetric approaches
nique in detecting and removing outliers in the neighborhood of show that the DMNE method outperforms classical PCA as well
each point of interest, resulting in improved normal estimation as more advanced methods such as MVE, MCD and MLESAC. Com-
near geometric discontinuities by incorporating only neighboring pared with conventional PCA-based normal estimation, DMNE
14 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

Fig. 13. Segmentation of the S1 model.

Fig. 14. Segmentation of the PCL dataset.


Table 4
Comparison of the conventional region growing against the presented segmentation method.

Models No. of ground truth segments Methods Metrics Total No. of segments
Precision (%) Recall (%) F1-score (%)
S1 13 PCL-RG (15°,95th) 100.0 57.1 72.7 7
PCL-RG (35°,98th) 100.0 25.0 40.0 4
Presented Method (10°) 92.3 100.0 96.0 13
T1 40 PCL-RG (15°,95th) 25.3 67.9 36.9 84
PCL-RG (35°,98th) 35.7 42.9 39.0 62
Presented Method (10°) 70.5 88.6 78.5 48
T2 34 PCL-RG (15°,95th) 3.3 100.0 6.4 392
PCL-RG (35°,98th) 14.1 95.0 24.5 136
Presented Method (10°) 38.9 100.0 56.0 54
PCL dataset 85 PCL-RG (3°,95th) 55.9 100.0 71.1 136
PCL-RG (5°,98th) 100.0 49.3 66.0 72
Presented Method (4°) 78.4 100.0 87.9 98
Table 49 PCL-RG (3°,95th) 51.3 100.0 67.8 78
PCL-RG (5°,98th) 48.1 64.1 54.9 66
Presented Method (4°) 90.2 100.0 94.8 51
A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16 15

requires more computational time, though computation time is not [13] H. Hoppe, T. DeRose, T. Duchamp, J. McDonald, W. Stuetzle, Surface
reconstruction from unorganized points, in: Proceedings of the 19th Annual
often a critical issue for point cloud segmentation.
Conference on Computer Graphics and Interactive Techniques, SIGGRAPH ’92,
In addition, a new region growing algorithm for the segmenta- ACM, New York, NY, USA, 1992, pp. 71–78, http://dx.doi.org/10.1145/
tion of unordered point clouds was also introduced. This approach 133994.134011.
employs the robust normal vectors and other salient local features [14] I. Jolliffe, Principal component analysis, in: Wiley StatsRef: Statistics Reference
Online, John Wiley & Sons, Ltd, 2002.
estimated using the DMNE process. Due to the higher accuracy in [15] N.J. Mitra, A. Nguyen, L. Guibas, Estimating surface normals in noisy point
local tangent plane fitting and higher distinctiveness of point nor- cloud data, Int. J. Comput. Geomet. Appl. 14 (04n05) (2004) 261–276, http://
mals, the segmentation algorithm needs only a single user-defined dx.doi.org/10.1142/S0218195904001470.
[16] Y. Wang, H.-Y. Feng, F.-t. Delorme, S. Engin, An adaptive normal
parameter to control the segmentation procedure. This is an estimation method for scanned point clouds with sharp features,
improvement over conventional methods which suffer from multi- Comput. Aided Des. 45 (11) (2013) 1333–1348, http://dx.doi.org/10.1016/
ple ambiguous input parameters to provide a trade-off between j.cad.2013.06.003.
[17] M. Pauly, M. Gross, L.P. Kobbelt, Efficient simplification of point-sampled
under- and over-segmentation. surfaces, in: Proceedings of the Conference on Visualization ’02, VIS ’02, IEEE
By utilizing a locally adaptive spatial connectivity analysis Computer Society, Washington, DC, USA, 2002, pp. 163–170.
within each segment growth step, it is possible to reach an appro- [18] E. Castillo, J. Liang, H. Zhao, Point Cloud segmentation and denoising via
constrained nonlinear least squares normal estimates, in: M. Breu, A.
priate segmentation solely based on geometric features. Conse- Bruckstein, P. Maragos (Eds.), Innovations for Shape Analysis, Mathematics
quently, the presented method is not tailored for a specific 3D and Visualization, Springer, Berlin, Heidelberg, 2013, pp. 283–299.
dataset or context and has been validated on a range of validation [19] B. Li, R. Schnabel, R. Klein, Z. Cheng, G. Dang, S. Jin, Robust normal estimation
for point clouds with sharp features, Comput. Graph. 34 (2) (2010) 94–106,
models to illustrate the flexibility for a variety of infrastructure
http://dx.doi.org/10.1016/j.cag.2010.01.004.
applications. [20] M.A. Fischler, R.C. Bolles, Random sample consensus: a paradigm for model
Future research can focus on optimizing and performing the fitting with applications to image analysis and automated cartography,
entire normal estimation and point cloud segmentation through Commun. ACM 24 (6) (1981) 381–395.
[21] K. Klasing, D. Althoff, D. Wollherr, M. Buss, Comparison of surface normal
parallel processing schemes utilizing the Graphics Processing Unit estimation methods for range sensing applications, in: IEEE International
(GPU) to increase the performance for massive 3D models. Conference on Robotics and Automation, 2009. ICRA ’09, 2009, pp. 3206–3211,
http://dx.doi.org/10.1109/ROBOT.2009.5152493.
[22] T. Song, F. Xi, S. Guo, Z. Ming, Y. Lin, A comparison study of algorithms for
surface normal determination based on point cloud data, Precis. Eng. 39 (2015)
Acknowledgments 47–55, http://dx.doi.org/10.1016/j.precisioneng.2014.07.005.
[23] M.M. Torok, M. Golparvar-Fard, K.B. Kochersberger, Image-based automated
This material is based upon the work supported by the National 3d crack detection for post-disaster building assessment, J. Comput. Civil Eng.
28 (5) (2013) A4014004.
Science Foundation (NSF) under Grant No. CMMI-1433765. The
[24] D. Lague, N. Brodu, J. Leroux, Accurate 3d comparison of complex topography
authors would also like to thank Prof. Mani Golparvar-Fard from with terrestrial laser scanner: application to the rangitikei canyon (nz), ISPRS J.
University of Illinois at Urbana-Champaign for graciously provid- Photogramm. Remote Sens. 82 (2013) 10–26.
[25] A.K. Patil, P. Holi, S.K. Lee, Y.H. Chai, An adaptive approach for the
ing S1, T1 and T2 point cloud datasets and their corresponding
reconstruction and modeling of as-built 3d pipelines from point clouds,
ground truth segmentation used in this research. Any opinions, Automat. Construct 75 (2017) 65–78.
findings, and conclusions or recommendations expressed in this [26] T. Czerniawski, M. Nahangi, C. Haas, S. Walbridge, Pipe spool recognition in
publication are those of the authors and do not necessarily reflect cluttered point clouds using a curvature-based shape descriptor, Automat.
Construct. 71 (2016) 346–358.
the views of the NSF. [27] R. Schnabel, R. Wahl, R. Klein, Efficient RANSAC for point-cloud shape
detection, Comput. Graph. Forum 26 (2) (2007) 214–226, http://dx.doi.org/
10.1111/j.1467-8659.2007.01016.x.
References [28] S. Filin, Surface clustering from airborne laser scanning data, Int. Arch.
Photogramm. Remote Sens. Spatial Inf. Sci. 34 (3/A) (2002) 119–124.
[29] P.J. Besl, R.C. Jain, Segmentation through variable-order surface fitting, IEEE
[1] R. Szeliski, Computer Vision: Algorithms and Applications, Springer Science &
Trans. Pattern Anal. Mach. Intell. 10 (2) (1988) 167–192.
Business Media, 2010.
[30] F. Tarsha-Kurdi, T. Landes, P. Grussenmeyer, et al., Hough-transform and
[2] M. Cabaleiro, B. Riveiro, P. Arias, J. Caamaño, Algorithm for beam deformation
extended ransac algorithms for automatic detection of 3d building roof planes
modeling from lidar data, Measurement 76 (2015) 20–31.
from lidar data, Proceedings of the ISPRS Workshop on Laser Scanning, vol. 36,
[3] B. Jafari, A. Khaloo, D. Lattanzi, Tracking structural deformations via automated
2007, pp. 407–412.
sample-based point cloud analysis, in: International Workshop on Computing
[31] M.F. Ahmed, C.T. Haas, R. Haas, Automatic detection of cylindrical objects in
in Civil Engineering 2017, ASCE, 2017.
built facilities, J. Comput. Civil Eng. 28 (3) (2014) 04014009.
[4] H. Son, C. Kim, C. Kim, 3d reconstruction of as-built industrial instrumentation
[32] G. Zhang, P.A. Vela, P. Karasev, I. Brilakis, A sparsity-inducing optimization-
models from laser-scan data and a 3d CAD database based on prior knowledge,
based algorithm for planar patches extraction from noisy point-cloud data,
Automat. Construct. 49 (Part B) (2015) 193–200, http://dx.doi.org/10.1016/j.
Comput.-Aided Civil Infrastruct. Eng. 30 (2) (2015) 85–102.
autcon.2014.08.007.
[33] J.M. Biosca, J.L. Lerma, Unsupervised robust planar segmentation of terrestrial
[5] A. Dimitrov, R. Gu, M. Golparvar-Fard, Non-uniform b-spline surface fitting
laser scanner point clouds based on fuzzy clustering methods, ISPRS J.
from unordered 3d point clouds for as-built modeling, Comput.-Aided Civil
Photogramm. Remote Sens. 63 (1) (2008) 84–98.
Infrastruct. Eng. 31 (7) (2016) 483–498, http://dx.doi.org/10.1111/mice.12192.
[34] T. Rabbani, F. Van Den Heuvel, G. Vosselmann, Segmentation of point clouds
[6] J. Jung, S. Hong, S. Yoon, J. Kim, J. Heo, Automated 3d wireframe modeling of
using smoothness constraint, Int. Arch. Photogramm. Remote Sens. Spatial Inf.
indoor structures from point clouds using constrained least-squares
Sci. 36 (5) (2006) 248–253.
adjustment for as-built bim, J. Comput. Civil Eng. (2015) 04015074.
[35] H. Son, C. Kim, C. Kim, Fully automated as-built 3d pipeline extraction method
[7] T. Gao, S. Ergan, B. Akinci, J. Garrett, Evaluation of different features for
from laser-scanned data based on curvature computation, J. Comput. Civil Eng.
matching point clouds to building information models, J. Comput. Civil Eng. 30
29 (4) (2014) B4014003.
(1) (2014) 04014107.
[36] S.B. Walsh, D.J. Borello, B. Guldur, J.F. Hajjar, Data processing of point clouds
[8] V. Pătrăucean, I. Armeni, M. Nahangi, J. Yeung, I. Brilakis, C. Haas, State of
for object detection for structural engineering applications, Comput.-Aided
research in automatic as-built modelling, Adv. Eng. Inform. 29 (2) (2015) 162–
Civil Infrastruct. Eng. 28 (7) (2013) 495–508, http://dx.doi.org/10.1111/
171.
mice.12016.
[9] A. Dimitrov, M. Golparvar-Fard, Segmentation of building point cloud models
[37] M. Yoon, Y. Lee, S. Lee, I. Ivrissimtzis, H.-P. Seidel, Surface and normal
including detailed architectural/structural features and mep systems,
ensembles for surface reconstruction, Comput. Aided Des. 39 (5) (2007) 408–
Automat. Construct. 51 (2015) 32–45.
420.
[10] H. Son, F. Bosché, C. Kim, As-built data acquisition and its use in production
[38] PCL, Region growing segmentation, 2016. <http://pointclouds.
monitoring and automated layout of civil infrastructure: a survey, Adv. Eng.
org/documentation/tutorials/region_growing_segmentation.php>.
Inform. 29 (2) (2015) 172–183.
[39] J.H. Friedman, J.L. Bentley, R.A. Finkel, An algorithm for finding best matches in
[11] A. Boulch, R. Marlet, Fast and robust normal estimation for point clouds with
logarithmic expected time, ACM Trans. Math. Software (TOMS) 3 (3) (1977)
sharp features, Comput. Graph. Forum 31 (5) (2012) 1765–1774, http://dx.doi.
209–226.
org/10.1111/j.1467-8659.2012.03181.x.
[40] G.H. Golub, C. Reinsch, Singular value decomposition and least squares
[12] T.K. Dey, G. Li, J. Sun, Normal estimation for point clouds: a comparison study
solutions, Numer. Math. 14 (5) (1970) 403–420, http://dx.doi.org/10.1007/
for a voronoi based method, in: Point-Based Graphics, 2005. Eurographics/IEEE
BF02163027.
VGTC Symposium Proceedings, IEEE, 2005, pp. 39–46.
16 A. Khaloo, D. Lattanzi / Advanced Engineering Informatics 34 (2017) 1–16

[41] P.C. Mahalanobis, On the generalized distance in statistics, Proc. Natl. Inst. Sci. [51] P.J. Rousseeuw, K.V. Driessen, A fast algorithm for the minimum covariance
(Calcutta) 2 (1936) 49–55. determinant estimator, Technometrics 41 (3) (1999) 212–223, http://dx.doi.
[42] P.J. Rousseeuw, A.M. Leroy, Robust regression and outlier detection, vol. 589, org/10.1080/00401706.1999.10485670.
John Wiley & Sons, 2005. [52] S. Choi, T. Kim, W. Yu, Performance evaluation of ransac family, J. Comput.
[43] V.J. Yohai, High breakdown-point and high efficiency robust estimates for Vision 24 (3) (1997) 271–300.
regression, Ann. Stat. 15 (2) (1987) 642–656. [53] P.H.S. Torr, A. Zisserman, MLESAC: a new robust estimator with application to
[44] K.S. Tatsuoka, D.E. Tyler, On the uniqueness of s-functionals and m-functionals estimating image geometry, Comput. Vis. Image Underst. 78 (1) (2000) 138–
under nonelliptical distributions, Ann. Stat. 28 (4) (2000) 1219–1243. 156, http://dx.doi.org/10.1006/cviu.1999.0832.
[45] P. Rousseeuw, V. Yohai, Robust regression by means of s-estimators, in: Robust [54] A. Nurunnabi, D. Belton, G. West, Robust segmentation in laser scanning 3d
and Nonlinear Time Series Analysis, Springer, 1984, pp. 256–272. point cloud data, in: 2012 International Conference on Digital Image
[46] M. Hubert, P. Rousseeuw, D. Vanpaemel, T. Verdonck, The DetS and DetMM Computing Techniques and Applications (DICTA), 2012, pp. 1–8, http://dx.
estimators for multivariate location and scatter, Comput. Stat. Data Anal. 81 doi.org/10.1109/DICTA.2012.6411672.
(2015) 64–75, http://dx.doi.org/10.1016/j.csda.2014.07.013. [55] P.J. Rousseeuw, C. Croux, Alternatives to the median absolute deviation, J. Am.
[47] N. Schertler, B. Savchynskyy, S. Gumhold, Towards globally optimal normal Stat. Assoc. 88 (424) (1993) 1273–1283.
orientations for large point clouds, Comput.Graph. Forum (2016), http://dx. [56] A. Khaloo, D. Lattanzi, Hierarchical dense structure-from-motion
doi.org/10.1111/cgf.12795, n/a–n/a. reconstructions for infrastructure condition assessment, J. Comput. Civil Eng.
[48] H.P. Lopuhaä, Multivariate s-estimators for location and scatter, Can. J. Stat. 19 (2016) 04016047.
(3) (1991) 307–321. [57] M. Awrangjeb, C.S. Fraser, An automatic and threshold-free performance
[49] S. Van Aelst, P. Rousseeuw, Minimum volume ellipsoid, Wiley Interdiscip. Rev.: evaluation system for building extraction techniques from airborne lidar data,
Comput. Stat. 1 (1) (2009) 71–82, http://dx.doi.org/10.1002/wics.19. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 7 (10) (2014) 4184–4198.
[50] R.A. Maronna, V.J. Yohai, Robust and efficient estimation of multivariate
scatter and location, Comput. Stat. Data Anal. 109 (2017) 64–75.

You might also like