Next Article in Journal
Molecularly Imprinted Polymer Based Sensors for Medical Applications
Next Article in Special Issue
Variational Bayesian Based Adaptive Shifted Rayleigh Filter for Bearings-Only Tracking in Clutters
Previous Article in Journal
Internet of Vehicles and Cost-Effective Traffic Signal Control
Previous Article in Special Issue
Collision Detection and Identification on Robot Manipulators Based on Vibration Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detail Preserved Surface Reconstruction from Point Cloud

1
National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
2
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(6), 1278; https://doi.org/10.3390/s19061278
Submission received: 18 February 2019 / Revised: 9 March 2019 / Accepted: 11 March 2019 / Published: 13 March 2019
(This article belongs to the Special Issue Sensor Signal and Information Processing II)

Abstract

:
In this paper, we put forward a new method for surface reconstruction from image-based point clouds. In particular, we introduce a new visibility model for each line of sight to preserve scene details without decreasing the noise filtering ability. To make the proposed method suitable for point clouds with heavy noise, we introduce a new likelihood energy term to the total energy of the binary labeling problem of Delaunay tetrahedra, and we give its s-t graph implementation. Besides, we further improve the performance of the proposed method with the dense visibility technique, which helps to keep the object edge sharp. The experimental result shows that the proposed method rivalled the state-of-the-art methods in terms of accuracy and completeness, and performed better with reference to detail preservation.

1. Introduction

Image-based scene reconstruction is a fundamental problem in Computer Vision. It has many practical applications in fields such as entertainment industry, robotics, cultural heritage digitalization and geographic systems. Image-based scene reconstruction has been studied for decades due to its low cost data acquisition and various usages. In recent years, researchers have made tremendous progress in this field. As far as small objects under controlled conditions are concerned, the performance of current scene reconstruction methods could achieve results comparable to those generated by laser scans or structured-light based methods [1,2]. However, when it comes to large scale scenes with multi-scale objects, current reconstruction methods have some problems with the completeness and accuracy, especially when concerning scene details [3].
Scene details such as small scale objects and object edges are an essential part of scene surfaces. Figure 1 shows an example of preserving scene details in reconstructing an ancient Chinese architecture. In general, cultural heritage digitalization projects, representing scene details such as the brackets in Figure 1, are among the most important tasks. The point cloud representation is often redundant and noisy, and the mesh representation is concise but it sometimes lose some information. Therefore, preserving scene details in reconstructing multi-scale scenes has been a difficult problem in surface reconstruction. The existing surface reconstruction methods [4,5,6,7] either ignore the scene details or rely on further refinement to restore them. Firstly, this is because, compared with noise, the supportive points in such part of the scene are sparse, making it difficult to distinguish true surface points from false ones. Secondly, the visibility models and associated parameters employed in existing methods are not particularly suitable for large scale ranges, where scene details are usually compromised for overall accuracy and completeness. While the first case seems to be unsolvable due to the lack of sufficient information, in this work, we focus on the second case. In particular, we extend the work of our conference paper [8], and suggest a new method with a new visibility model for surface reconstruction.
In many previous surface reconstruction methods [4,5,6,9,10,11,12,13], visibility information that records a 3D point is seen by the views used to help to generate accurate surface meshes. To use the visibility information, assumptions of the visibility model are made so that the space between camera center and the 3D point is free-space and the space behind the point along the line of sight is full-space. However, the points are often contaminated with noise and the full-space scales are often hard to determinate. To preserve scene details without decreasing the noise filtering ability, we propose a new visibility model with error tolerance and adaptive end weights. We also introduce a new likelihood energy representing the punishment of wrongly classifying a part of space as free-space or full-space, which helps to improve the ability of the proposed method to efficiently filter noise. Moreover, we further improve the performance of the proposed method with the dense visibility technique, which helps to keep the object edge sharp. Experimental results show that the proposed method rivals the state-of-the-art methods in terms of accuracy and completeness, and performs better with reference to detail preservation.

2. Related Work

In recent years, various works have been done to advance the image-based scene reconstruction. Referring to small objects, silhouette based methods [14,15,16,17,18] are proposed. The silhouettes provide proper bounds for the objects, which help reduce the computing cost and yield a good model for the scene. However, good silhouettes rely on effective image segmentation, which remains a difficult task. Furthermore, silhouettes can hardly be used for large scale scenes. Volumetric methods such as space carving [19,20,21], level sets [22,23] and volumetric graph cut [9,10,11] often yield good results for small objects. However, in the case of large scale scenes, the computational and memory costs increase rapidly as scene scale grows. Consequently, this makes them unsuitable for large scale scene reconstruction. Vogiatzis et al. [9] proposed a volumetric graph cut based method to reconstruct an object by labeling voxels as inside or outside, in which a photo-consistency term is introduced to enhance the final result. Tran and Davis [10] and Lempitsky et al. [11] also exploited the same idea, the former by adding predetermined locations of possible surface as surface constraints and the latter by estimating visibility based on position and orientation of local surface patches, then optimizing on a CW-complex.
As far as outdoor scenes, uncontrollable imaging conditions and multiple scale structures make it hard to reconstruct scene surfaces. A common process of reconstructing large scale scenes is to generate the dense point cloud from calibrated images first, and then extract the scene surface. The dense point cloud can be generated through the depth fusion method [24,25,26,27,28] or through the feature expansion methods [29]. In depth-map fusion methods, depth-maps are usually computed independently, and then merged into one point cloud; in feature expansion methods instead, the sparse point cloud is generated first, and then expanded to points near the seed points. Usually, the depth-map fusion based methods yield a relatively denser but noisier point cloud. Once the dense point cloud is generated, scene surface can be reconstructed through poisson surface reconstruction [30] or through graph cut based methods [4,5,12,13]. In [12], firstly the dense point cloud is used to generate Delaunay tetrahedra, then a visibility model is introduced to weight the facets in Delaunay tetrahedra, and finally the inside–outside binary labeling of tetrahedra is solved and the surface is extracted; it consists of triangles between tetrahedra with different labels. The basic assumption of the visibility model in [12] is that the space between camera center and 3D point is free-space, while the space behind 3D point is full-space. Then, this typical visibility model is promoted by a refined version in [13], namely soft-visibility, to cope with noisy point clouds. Apart from using dense point clouds, Bódis-Szomorú et al. [31] also proposed a method to produce a surface mesh by fitting the meshes reconstructed by single views and a sparse point cloud. The method is parallelizable and the quality of the final meshes is comparable to that of a state-of-the-art pipeline [32].
Besides the point-cloud based methods, large scale scene reconstruction can also be achieved through volume-based methods. Häne et al. [6,33] proposed a method for scene reconstruction and object classification. The scene space is represented by voxels and each one is given a class label. The accuracy of the classification relies on the decision tree, which is made up of labeled images. The extensive works in [34,35] reduce the heavy memory cost of the method in [33] by introducing octree structure and block scheme, respectively. Savinov et al. [36] exploited an idea similar to that in [6,33], and used full multilabel ray potential and continuously inspired anisotropic surface regularization together to yield 3D semantic models. Ummenhofer and Brox [37] proposed a method that is capable of handling a billion points. This method uses an octree structure to manage the points and then reconstruct the scene using a level-set alike method.
In this paper, we extend the work of our conference paper [8], follow the idea in [5,13] and exploit the visibility model for the line of sight in the binary labeling problem of Delaunay tetrahedra. The main contributions of the proposed method are: (1) two new visibility models for the visibility information of the points on the vertices of Delaunay tetrahedra and inside them, respectively; and (2) the dense visibility technique, which exploits the visibility information of unmerged points for better performance of preserving scene details. Experimental comparison results show that the proposed method rivals the state-of-the-art methods [5,24,27,29,38] in terms of accuracy and completeness, and performs better in detail preservation.

3. Visibility Models and Energies

The pipeline of the proposed method is shown in Figure 2. The input of the proposed method is a dense point cloud generated by multi-view stereo (MVS) methods. Each point in the point cloud is attached with the visibility information recording that from which views the point is seen. Next, Delaunay tetrahedra are constructed from given input point cloud, and the scene reconstruction problem is formulated as a binary labeling problem to label tetrahedra as inside or outside. Then, an s-t graph is constructed with tetrahedra as the vertices and facets of tetrahedra as the edges. Finally, by minimizing the energy defined on the s-t graph, the vertices are separated into two parts, i.e., inside and outside, and the scene surface is extracted which consists of triangle facets lying between tetrahedra with different labels. The key issue of this process is to find a proper energy. In doing so, we introduce a new visibility model to formulate the energy of the binary labeling problem. In the following subsections, we detail the visibility models in [12,13] and our new one. Before that, we give the meaning of the symbols used in this work in Table 1 for better understanding.

3.1. Existing Visibility Models

The typical visibility model [12] assumes that, for each line of sight v, the space between camera center c and point p is free-space; the space behind point p along the line of sight is full-space. For the Delaunay tetrahedra of the point cloud, the tetrahedra intersected by segment ( c , p ) should be labeled as outside; the tetrahedron right behind the point p should be labeled as inside. For example, in Figure 3a, according to the above model, tetrahedra T 1 T 5 should be labeled as outside, while tetrahedron T 6 should be labeled as inside.
For a single line of sight, the above label assignment is desirable. While taking all lines of sight into consideration, some surface part might not be handled properly, for example, that between T 2 and T 3 in Figure 3a. This scenario violates the label assignment principle described above. To punish the conflicts, the facets intersected by a line of sight are given weights (energy) of W ( l T i , l T j ) . Similarly, weights for punishing bad label assignments of the first tetrahedron and last tetrahedron are D ( l T 1 ) and D ( l T N v + 1 ) , respectively. Therefore, the visibility energy is the sum of the penalties of all the bad label assignments in all the lines of sight, as
E v i s t y p i c a l = v V [ D ( l T 1 ) + i = 1 N v 1 W ( l T i , l T i + 1 ) + D ( l T N v + 1 ) ]
where
D ( l T 1 ) = α v if l T 1 = 1 0 otherwise , D ( l T N v + 1 ) = α v if l T N v + 1 = 0 0 otherwise , W ( l T i , l T j ) = α v if l T i = 0 l T j = 1 0 otherwise
where V denotes the visibility information set, containing all lines of sight; N v is the number of tetrahedra intersected by a single line of sight v, indexed from the camera center c to the point p; N v + 1 denotes the tetrahedron behind the point p; l T is the label of tetrahedron T, with 1 stands for inside and 0 outside; α v is the weight of a line of sight v, which can be the photo consistency score of point p in current view.
The typical visibility model as well as the energy formulation in Equation (1) described above is effective in most cases, but it has several flaws in relation to dense and noisy point clouds [13]. The surfaces reconstructed by [12] tend to be overly complex, with bumps on the surface and handles inside the model. One possible solution to these problems is the soft visibility proposed in [13], as shown in Figure 3b. In the soft visibility model, the basic assumption is similar to the previous typical visibility model. However, the edge weights are multiplied by a weight factor ( 1 e d 2 / 2 σ 2 ) , in which d represents the distance between the point p and the intersecting point. In addition, the end tetrahedron in the soft visibility model is shifted to a distance of 3 σ along the line of sight.

3.2. Our Proposed Visibility Model

Although the soft visibility model is effective to filter noise points and helps to yield visually smoothed models, it sometimes performs poorly in preserving details, especially in a large scene containing some relatively small scale objects (see Experimental Results). According to our observations, this happens mainly because of the improperly chosen relaxation parameter σ and the strong constraint imposed on the end of line of sight in the tetrahedron k σ from the point p along the line of sight. In some cases, such end tetrahedra would be free-space even though the point p is a true surface point.
To balance noise filtering and detail preserving, we propose a new visibility model, which is shown in Figure 4a. In our visibility model, we also use the relaxed visibility constraints in the soft visibility model in the space between the camera center c and the point p, i.e., the weight factor ( 1 e d 2 / 2 σ 2 ) is kept to ensure that the final model is not overly complex. Then, we set the end of line of sight in the tetrahedron just right behind the point p, to avoid the wrong end in the soft visibility model in the case of small scale objects. To determine the weight of the t-edge of the end tetrahedron, we compare the end tetrahedra of noisy points and true surface points on datasets with quasi-truth. Figure 4b shows a typical end tetrahedron of noisy points and that of true surface points on densely sampled surfaces in 2D space. Noise points tend to appear in somewhere a bit away from true surface, which makes the end tetrahedra (triangles in 2 D ) thin and long, and true surface points are often surrounded by other true surface points, which makes their end tetrahedra flat and wide. Based on the above observations, we set a weight of α v ( 1 e r 2 / 2 σ 2 ) to the t-edge of the end tetrahedron, where r is the radius of the circumsphere of the end tetrahedron.
With our new visibility model, our visibility energy is formulated as
E v i s = v V [ D ( l T 1 ) + i = 1 N v 1 W ( l T i , l T i + 1 ) + D ( l T N v + 1 ) ]
where
D ( l T 1 ) = α v if l T 1 = 1 0 otherwise , D ( l T N v + 1 ) = α v ( 1 e r 2 / 2 σ 2 ) if l T N v + 1 = 0 0 otherwise ,
W ( l T i , l T j ) = α v ( 1 e d 2 / 2 σ 2 ) if l T i = 0 l T j = 1 0 otherwise
Instead of setting the tolerance parameter σ to a constant manually, we set σ adaptively within 0.5–1% of the length of line of sight in our visibility model. The underlying reason is that, typically, in depth-map fusion methods, the error bound used to filter bad correspondences while generating dense point clouds is adaptively set to 0.5–1% of the depth of the point in the current view, as in [28]. This gives each point a confidence interval along the line of sight.

4. Likelihood Energy for Efficient Noise Filtering

In both the typical visibility model and our proposed one, the end of each line of sight is set in the tetrahedra right behind the point p. This practice sometimes could weaken the ability of noise filtering. When the surface is sampled very densely, unexpected handles could appear inside the model [13], or part of the surface could fail to be reconstructed, as shown in Figure 5. This is mainly due to the unbalanced links of s-edges and t-edges in the s-t graph, i.e., the s-edges are too strong to be cut as the camera centers are consistent for all lines of sight, while the t-edges are weak because their weights are scattered by the varying locations of the points. The noisier the point cloud is, the greater the gap between s-edges and t-edges becomes.

4.1. Likelihood Energy

To solve these problems, we introduce a likelihood energy E l i k e to the total energy of the binary labeling problem. E l i k e is defined as
E l i k e = i = 1 N D ( l T i ) , where D ( l T i ) = U o u t ( T i ) if l T i = 1 U i n ( T i ) otherwise
where N is the total number of Delaunay tetrahedra. E l i k e measures the penalties of a wrong label assignment. For each tetrahedron, it is attached with two attributes that describe how likely it is to be outside or inside. If it is mistakenly labeled, a penalty is introduced, i.e., U o u t ( T i ) or U i n ( T i ) .
To evaluate the likelihood of the label assignment of a tetrahedron, we employ the measure free-space support [5], which is used to measure the emptiness of a compact space. For each line of sight v that intersects tetrahedra T, it contributes to the emptiness of T i , thus increasing the probability of T to be outside. The free-space support f ( T ) of tetrahedron T is computed as
f ( T ) = v V T α v , with V T = { v | v T }
where V T is the set of lines of sight v that intersect with tetrahedron T. To evaluate f ( T ) to adequately describe the likelihood energy, we set U o u t ( T ) = λ f ( T ) and U i n ( T ) = λ ( β f ( T ) ) , where λ is a constant used to scale the range of f ( T ) , and β is a constant greater than f ( T ) for all tetrahedra T.

4.2. Implementation of the Likelihood Energy

For the likelihood term E l i k e , we link two edges for each vertex i in s-t graph. One is from source s to vertex i, with weight U o u t ( T i ) ; the other is from vertex i to sink t, with weight U i n ( T i ) . However, to reduce the complexity of the graph, we cross out the s-edges of all vertices, and only link the vertices to sink t whose correspondent tetrahedra have lower f ( T ) than the 75th percentile of all f ( T ) s. Since some of the vertices in s-t graph have a heavy-weight edge linked to source s, we rely on the visibility constraints to label truly outside tetrahedra, instead of linking them with extra s-edges.
Note that the free space support threshold of 75th percentile was empirically set, as shown in Figure 6 and Figure 7. We generated dense point clouds of four datasets, and applyiedDelaunay tetrahedralization to them. The four datasets were Fountain-P11 [3], Herz-Jesu-P8 [3], Temple-P312 [1] and scan23-P49 [2]. Then, we labeled the tetrahedra with method described in [5], and took the result as the quasi-truth. Finally, we evaluated the ratios of number of outside and inside tetrahedra in different proportions of all free-space support scores, as shown in Figure 6. In Figure 6, we can see that, generally, tetrahedra with lower f ( T ) ( f ( T ) lower than the 75th percentile of all free-space support scores) had a higher probability to be truly inside. As shown in Figure 7, the true positive rates and false positive rates with different free-space support thresholds were evaluated. It is noteworthy that, when the free-space support threshold was set as the 75th percentile of each dataset, both true positive rate and false positive rate were reasonable. Therefore, we only linked those tetrahedra to sink t whose free-space support was lower than 75th percentile of all free-space support scores. In Figure 5, we can see that the likelihood energy is helpful for filtering noise.

5. Surface Reconstruction with Energy Minimization

With the likelihood energy and the proposed visibility model, the total energy of the binary labeling problem of the Delaunay tetrahedra is formulated as
E t o t a l = E v i s + λ l i k e E l i k e + λ q u a l E q u a l
where λ l i k e and λ q u a l are two constant balancing factors; E v i s and E l i k e are defined in Equations (2) and (3); and E q u a l is the surface quality energy introduced in [13] as
E q u a l = f w f
where
w f = 1 min { cos ( ϕ ) , cos ( ψ ) } , if l T 1 f l T 2 f
The total energy E t o t a l defined in Equation (5) could be represented as an s-t graph and minimized by the Maxflow/Mincut algorithm [39]. In E t o t a l , the graph construction for the visibility energy E v i s and the surface quality energy E q u a l is straightforward, and the likelihood E l i k e is implemented as described in Section 4.2. Then, the energy minimization problem is solved using the Maxflow/Mincut algorithm on the s-t graph, and the optimal label assignment of Delaunay tetrahedra is yielded. Ultimately, a triangle mesh, which consists of triangles lying between tetrahedra with different labels, is extracted. A further optional refinement can be applied as described in [7].

6. Dense Visibility for Edge Preservation

Although the proposed visibility model in Figure 4 and the energy formulation in Equation (5) are carefully designed for preserving the scene details, the object edges sometimes still appear to be inaccurate concerning bumps and dents. When referring to the original depth maps, the depths of object edges are quite smooth. This scenario is shown in Figure 8. Figure 8a,b shows the object edges in 3D meshes reconstructed by the method in [5] and the proposed method described in the previous sections, as well as the original image and the corresponding depth map with a similar view point. We can easily see that the object edges in the reconstructed meshes failed to keep the smoothness as in the depth map. This could be due to the error of either point locations or visibility information which is introduced in the depth map fusion process. To circumvent such problem, instead of fusing the depths in matched cameras, we generate the dense point cloud simply by joining all of the 3D points recovered with depth maps and camera parameters. This could result in a much denser and more redundant point cloud, but it also contains more useful information for better surface reconstruction. To alleviate the memory and the computational cost, points are sampled and then Delaunay tetrahedra are constructed from the sampled point cloud. Instead of discarding those points unused for tetrahedralization and their visibility information, we apply a modified version of our visibility model to use their visibility information, as shown in Figure 9. To keep the ability to filter noise and select true surface points, we keep most of the visibility model in Figure 4 along the line of sight and only modify the part near the 3D point. The difference between the visibility models in Figure 9 and the one in Figure 4 is that, for a point p that lies in a tetrahedron, the end of the line of sight is set in the tetrahedron right behind the one that contains p, and the facet between them is punished with a weight multiplied by the same weight factor as in the soft visibility model. In this way, we keep the end of the line of sight close to the 3D point and the tolerance to noise. In Figure 8c, we show the surface meshes containing object edges reconstructed by the method in [5] and the proposed method with the dense visibility technique. We can infer from the results in Figure 8 that the dense visibility technique is helpful for preserving object edges.

7. Experimental Results

In our experiments, the input dense point cloud was generated from images with the open source library OpenMVG (http://imagine.enpc.fr/~moulonp/openMVG/) and OpenMVS (http://cdcseacave.github.io/openMVS/). Sparse point cloud was generated by OpenMVG, then densified with OpenMVS. Delaunay tetrahedralization was computed using CGAL (http://www.cgal.org/) library. Maxflow/Mincut algorithm [39] aws used. The proposed method was tested on public benchmark MVS dataset [2] and Tanks and Temples dataset [40].
We first tested the proposed method on MVS dataset [2]. MVS dataset [2] contains over one hundred scenes consisting of images depicting compact objects under controlled lighting conditions. Figure 10 shows the result on the MVS dataset [2]. The reference model is in the first column. From the second column to the last column, there are the models of Tola et al. [27], Furukawa and Ponce [29], Campbell et al. [24], Jancosek and Pajdla [5] and the proposed method, respectively. The final meshes given by Tola et al. [27], Furukawa and Ponce [29] and Campbell et al. [24] were generated by Poisson surface reconstruction method [30] and trimmed, which were provided in the benchmark MVS dataset [2]; the meshes of Jancosek and Pajdla [5] were generated by OpenMVS, which contains an reimplementation of the method in [5]. Figure 10 shows that the proposed method could reconstruct complex scenes as well as regular scenes, and it had great potential for preserving scene details. Figure 11 shows the accuracy and completeness of reconstructed meshes over twenty scenes, which are scans 1–6, 9–10, 15, 21, 23–24, 29, 36, 44, 61, 110, 114, 118 and 122. The detailed information of the 3D models evaluated on MVS dataset [2] is presented in Table 2. The evaluation method is described in [2], in which accuracy is measured as the distance from the MVS reconstruction to the reference model, and the completeness is measured from the reference model to the MVS reconstruction. In addition to the evaluation of the surface, we also evaluated the point clouds generated by the methods in [24,27,29] as well as OpenMVS, with the point clouds in [24,27,29] provided by MVS dataset [2]. We can see in Figure 11 that generally the point clouds achieved better scores than the surface meshes, which complies with the results in [2]. The underlying reason could be that the mesh representation discarded points that are redundant but close to the true surface, and simultaneously fixed unexpected gaps. Comparing the result within the surface meshes, the proposed method without the dense visibility technique was not outstanding, in terms of both accuracy and completeness. By applying the dense visibility technique, the proposed method achieved the best median accuracy and median completeness, and the second best mean accuracy and mean completeness. In odd rows of Figure 12 are the local point cloud and the surface meshes of Jancosek and Pajdla [5], the proposed method without the dense visibility technique and the proposed method; in the even rows of Figure 12 are the evaluation result (lower is better) of the corresponding local models through the method in [41]. We show the ability of the three methods to preserve thin objects and object edges. In some cases, the method in [5] failed completely in reconstructing them; even the complete ones were less accurate, both visually and quantitatively, than those provided by the proposed method with or without the dense visibility technique. In addition, compared with the method in [5], the proposed method showed a tremendous capability of preserving sharp object edges.
To better visualize the differences between the models generated by Jancosek and Pajdla [5] and the proposed method, we enlarged some parts of the meshes generated by the two methods. The enlarged views are shown in Figure 12. We also tested the proposed method on benchmark Tanks and Temples dataset [40]. Scenes in Tanks and Temples dataset [40] are realistic and contain plenty of objects with different scales in both outdoor conditions and indoor environments. We evaluated the proposed method as well as two other methods on four outdoor scenes (Barn, Ignatius, Truck and Courthouse) of the training set and the eight scenes of the intermediate set of Tanks and Temples dataset [40]. Figure 13 shows the result of the proposed method on four scenes of the training set of Tanks and Temples dataset [40]. Looking at Figure 13 from left to right, we can see the input images, the precision and recall of the model generated by the proposed method and the F-scores of the models generated by Colmap [38], Jancosek and Pajdla [5], and the proposed method with and without the dense visibility technique. The evaluation method is described in [40], in which precision is measured as the distance from the MVS reconstruction to the reference model, the completeness is measured from the reference model to the MVS reconstruction, and the F-score is the harmonic mean of precision and recall with a given threshold. Table 3 presents the evaluation result of the 3D models through the method in [42]. Since the evaluation method in [40,42] takes point clouds as the input, the meshes generated by Jancosek and Pajdla [5] and the proposed method were sampled to acquire point clouds. The point clouds yielded by Colmap [38] are provided in Tanks and Temples dataset [40]. From the evaluation result of the four scenes in Figure 13 and Table 3, we concluded similarly as for the MVS dataset [2] that the proposed method outperformed the method in [5] with the dense visibility technique, while it performd slightly worse than the method in [5] without it. The proposed method performed better than the other three methods in Barn, Truck and Courthouse and rivalled Colmap [38] in Ignatius when evaluated through the method in [40], while it achieved the best result in the four scenes when evaluated through the method in [42]. Figure 14 shows the detailed views of the proposed method and the method of Jancosek and Pajdla [5] on four scenes of the training set. As in Figure 12, we also present the local models and the evaluation result using the method in [41] in odd rows and even rows, respectively. In Figure 14, Jancosek and Pajdla’s method [5] inaccurately reconstructed the edge of the roof, wrongly fixed the gap between the arm and the chest of the statue, and failed to reconstruct the rearview mirror of the truck and the light stand of the lamp, while the proposed method performed well in these parts. It is noteworthy that the proposed method also had a great ability of deducing the close form of the real surface when there were no supportive points due to the occlusion, a common phenomenon in 3D reconstruction. A good example is the arm and the chest of the statue in Figure 14. We can see that the proposed method fixed the vacant part of the chest reasonably, even though it was occluded by the arm of the statue, while the method in [5] gave a less satisfactory solution. The underlying reason is that, in the proposed visibility models, the end of a line of sight lies in the tetrahedra right behind the 3D point, which is good for separating small scale objects from other objects. Table 4 and Figure 15 present the result of the proposed method on the intermediate set of Tanks and Temples dataset [40]. The detailed information of the 3D models evaluated on the intermediate set of Tanks and Temples dataset [40] is presented in Table 5. In Table 4, we can see that the result of the proposed method and that in [5] are not outstanding. However, compared to the evaluation result of the point clouds of OpenMVG + OpenMVS, the result of both the proposed method and the method in [5] achieves a substantial boost in all scenes except M60. The underlying reason could be that the two methods simultaneously filtered most noise points and fixed unexpected gaps during the surface reconstruction process. These two behaviors had opposite effects on the evaluation result, since the former one increased the F-score while the latter one decreased it. In Table 4, we can also find that generally with the dense visibility technique the proposed method outperformed the method in [5], while without the dense visibility technique it performed worse than the method in [5]. Figure 15 shows the detailed views of the proposed method and the method of Jancosek and Pajdla [5] on eight scenes of the intermediate set. Table 4 shows the F-scores of several public methods and the proposed method on eight scenes of the intermediate set. In dealing with the thin objects such as the human legs in Francis, the horse ear in Horse, the wire in M60, the barrel in Panther, the bar in Playground and the handle in Train, Jancosek and Pajdla’s method [5] either failed to reconstruct it or wrongly fixed the gaps, while the proposed method performed well in these parts. Compared to the meshes of the method in [5], the proposed method also kept distinguishing object edges such as the cloth folds in Family, the horse mouth in Horse and the doorframe in Lighthouse. Therefore, we can infer that the proposed method has a strong ability to preserve the scene details.

8. Conclusions

In this paper, we present a new surface reconstruction method. The proposed method is designed to preserve scene details while keeping the ability to filter noise. To make the proposed method efficient to filter out noise and to select true surface points, we introduce a new visibility model with error tolerance and adaptive end weights. Along with the proposed visibility model, a new likelihood energy term is added to the total energy of the binary labeling problem to promote the robustness of the proposed method to noise. Moreover, we further improve the performance of the proposed method with the dense visibility technique, which avoids the error introduced in the point cloud generation process and provide denser visibility information. We tested the proposed method on two publicly available benchmark datasets. Experimental results on different datasets show that the proposed method rivalled the state-of-the-art methods in terms of accuracy and completeness, while it could preserve scene details such as thin objects and sharp edges. Our future work will consist in segmenting the model and adding semantic knowledge to each part of the model.

Author Contributions

Conceptualization, Y.Z.; Data curation, Y.Z. and S.S.; Formal analysis, Y.Z.; Funding acquisition, S.S. and Z.H.; Investigation, Y.Z.; Methodology, Y.Z. and S.S.; Project administration, S.S. and Z.H.; Resources, Y.Z. and S.S.; Software, Y.Z. and S.S.; Supervision, S.S. and Z.H.; Validation, Y.Z.; Visualization, Y.Z.; Writing—original draft, Y.Z.; and Writing—review and editing, Y.Z., S.S. and Z.H.

Funding

This research was funded by National Natural Science Foundation of China grant numbers 61632003, 61873265 and 61573351.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Seitz, S.M.; Curless, B.; Diebel, J.; Scharstein, D.; Szeliski, R. A comparison and evaluation of multi-view stereo reconstruction algorithms. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition IEEE, New York, NY, USA, 17–22 June 2006; pp. 519–528. [Google Scholar]
  2. Jensen, R.; Dahl, A.; Vogiatzis, G.; Tola, E.; Aanæs, H. Large scale multi-view stereopsis evaluation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition IEEE, Columbus, OH, USA, 23–28 June 2014; pp. 406–413. [Google Scholar]
  3. Strecha, C.; von Hansen, W.; Van Gool, L.; Fua, P.; Thoennessen, U. On benchmarking camera calibration and multi-view stereo for high resolution imagery. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition IEEE, Anchorage, AL, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  4. Sinha, S.N.; Mordohai, P.; Pollefeys, M. Multi-view stereo via graph cuts on the dual of an adaptive tetrahedral mesh. In Proceedings of the 11th IEEE International Conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  5. Jancosek, M.; Pajdla, T. Exploiting visibility information in surface reconstruction to preserve weakly supported surfaces. Int. Schol. Res. Not. 2014, 2014, 798595. [Google Scholar] [CrossRef] [PubMed]
  6. Häne, C.; Zach, C.; Cohen, A.; Pollefeys, M. Dense semantic 3d reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1730–1743. [Google Scholar] [CrossRef] [PubMed]
  7. Vu, H.H.; Labatut, P.; Pons, J.P.; Keriven, R. High accuracy and visibility-consistent dense multiview stereo. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 889–901. [Google Scholar] [CrossRef] [PubMed]
  8. Zhou, Y.; Shen, S.; Hu, Z. A New Visibility Model for Surface Reconstruction. In Proceedings of the CCF Chinese Conference on Computer Vision, Tianjin, China, 11–14 October 2017; pp. 145–156. [Google Scholar]
  9. Vogiatzis, G.; Torr, P.H.; Cipolla, R. Multi-view stereo via volumetric graph-cuts. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE, San Diego, CA, USA, 20–25 June 2005; pp. 391–398. [Google Scholar]
  10. Tran, S.; Davis, L. 3D surface reconstruction using graph cuts with surface constraints. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 219–231. [Google Scholar]
  11. Lempitsky, V.; Boykov, Y.; Ivanov, D. Oriented visibility for multiview reconstruction. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 226–238. [Google Scholar]
  12. Labatut, P.; Pons, J.P.; Keriven, R. Efficient multi-view reconstruction of large-scale scenes using interest points, delaunay triangulation and graph cuts. In Proceedings of the 11th IEEE international Conference on Computer Vision, Rio De Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  13. Labatut, P.; Pons, J.P.; Keriven, R. Robust and efficient surface reconstruction from range data. Comp. Graph Forum 2009, 28, 2275–2290. [Google Scholar] [CrossRef]
  14. Laurentini, A. The visual hull concept for silhouette-based image understanding. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 150–162. [Google Scholar] [CrossRef]
  15. Esteban, C.H.; Schmitt, F. Silhouette and stereo fusion for 3D object modeling. Comp. Vis. Image Underst. 2004, 96, 367–392. [Google Scholar] [CrossRef]
  16. Starck, J.; Miller, G.; Hilton, A. Volumetric stereo with silhouette and feature constraints. In Proceedings of the British Machine Vision Conference, Edinburgh, UK, 4–7 September 2006; pp. 1189–1198. [Google Scholar]
  17. Franco, J.S.; Boyer, E. Fusion of multiview silhouette cues using a space occupancy grid. In Proceedings of the 10th IEEE International Conference on Computer Vision, San Diego, CA, USA, 20–25 June 2005; pp. 1747–1753. [Google Scholar]
  18. Guan, L.; Franco, J.S.; Pollefeys, M. 3d occlusion inference from silhouette cues. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  19. Kutulakos, K.N.; Seitz, S.M. A theory of shape by space carving. Int. J. Comp. Vis. 2000, 38, 199–218. [Google Scholar] [CrossRef]
  20. Broadhurst, A.; Drummond, T.W.; Cipolla, R. A probabilistic framework for space carving. In Proceedings of the 8th IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 388–393. [Google Scholar]
  21. Yang, R.; Pollefeys, M.; Welch, G. Dealing with Textureless Regions and Specular Highlights-A Progressive Space Carving Scheme Using a Novel Photo-consistency Measure. In Proceedings of the 9th IEEE International Conference on Computer Vision, IEEE, Nice, France, 13–16 October 2003; pp. 576–584. [Google Scholar]
  22. Jin, H.; Soatto, S.; Yezzi, A.J. Multi-view stereo reconstruction of dense shape and complex appearance. Int. J. Comp. Vis. 2005, 63, 175–189. [Google Scholar] [CrossRef]
  23. Pons, J.P.; Keriven, R.; Faugeras, O. Multi-view stereo reconstruction and scene flow estimation with a global image-based matching score. Int. J. Comp. Vis. 2007, 72, 179–193. [Google Scholar] [CrossRef]
  24. Campbell, N.D.; Vogiatzis, G.; Hernández, C.; Cipolla, R. Using multiple hypotheses to improve depth-maps for multi-view stereo. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 766–779. [Google Scholar]
  25. Goesele, M.; Curless, B.; Seitz, S.M. Multi-view stereo revisited. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006; pp. 2402–2409. [Google Scholar]
  26. Hiep, V.H.; Keriven, R.; Labatut, P.; Pons, J.P. Towards high-resolution large-scale multi-view stereo. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1430–1437. [Google Scholar]
  27. Tola, E.; Strecha, C.; Fua, P. Efficient large-scale multi-view stereo for ultra high-resolution image sets. Mach. Vis. Appl. 2012, 23, 903–920. [Google Scholar] [CrossRef]
  28. Shen, S. Accurate multiple view 3d reconstruction using patch-based stereo for large-scale scenes. IEEE Trans. Image Process. 2013, 22, 1901–1914. [Google Scholar] [CrossRef] [PubMed]
  29. Furukawa, Y.; Ponce, J. Accurate, dense, and robust multiview stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1362–1376. [Google Scholar] [CrossRef] [PubMed]
  30. Kazhdan, M.; Hoppe, H. Screened poisson surface reconstruction. ACM Trans. Graph. 2013, 32, 29. [Google Scholar] [CrossRef]
  31. Bódis-Szomorú, A.; Riemenschneider, H.; Van Gool, L. Superpixel meshes for fast edge-preserving surface reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2011–2020. [Google Scholar]
  32. Jancosek, M.; Pajdla, T. Multi-view reconstruction preserving weakly-supported surfaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 3121–3128. [Google Scholar]
  33. Hane, C.; Zach, C.; Cohen, A.; Angst, R.; Pollefeys, M. Joint 3D scene reconstruction and class segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 201; pp. 97–104.
  34. Blaha, M.; Vogel, C.; Richard, A.; Wegner, J.D.; Pock, T.; Schindler, K. Large-scale semantic 3d reconstruction: an adaptive multi-resolution model for multi-class volumetric labeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, Las Vegas, NE, USA, 26 June–1 July 2016; pp. 3176–3184. [Google Scholar]
  35. Cherabier, I.; Hane, C.; Oswald, M.R.; Pollefeys, M. Multi-label semantic 3d reconstruction using voxel blocks. In Proceedings of the 4th International Conference on 3D Vision, Stanford University, Stanford, CA, USA, 25–28 October 2016; pp. 601–610. [Google Scholar]
  36. Savinov, N.; Hane, C.; Ladicky, L.; Pollefeys, M. Semantic 3d reconstruction with continuous regularization and ray potentials using a visibility consistency constraint. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NE, USA, 27–30 June 2016; pp. 5460–5469. [Google Scholar]
  37. Ummenhofer, B.; Brox, T. Global, dense multiscale reconstruction for a billion points. In Proceedings of the IEEE International Conference on Computer Vision, IEEE, Washington, DC, USA, 7–13 December 2015; pp. 1341–1349. [Google Scholar]
  38. Schönberger, J.L.; Frahm, J.M. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NE, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar]
  39. Goldberg, A.V.; Hed, S.; Kaplan, H.; Tarjan, R.E.; Werneck, R.F. Maximum flows by incremental breadth-first search. In Proceedings of the European Symposium on Algorithms, Saarbrücken, Germany, 5–9 September 2011; pp. 457–468. [Google Scholar]
  40. Knapitsch, A.; Park, J.; Zhou, Q.Y.; Koltun, V. Tanks and Temples: Benchmarking Large-Scale Scene Reconstruction. ACM Trans. Graph. 2017, 36, 78. [Google Scholar] [CrossRef]
  41. Jurjević, L.; Gašparović, M. 3D Data Acquisition Based on OpenCV for Close-range Photogrammetry Applications. In Proceedings of the ISPRS Hannover Workshop: HRIGI 17–CMRT 17–ISA 17–EuroCOW 17, Hannover, Germany, 6–9 June 2017. [Google Scholar]
  42. Sapirstein, P. Accurate measurement with photogrammetry at large sites. J. Archaeol. Sci. 2016, 66, 137–145. [Google Scholar] [CrossRef]
Figure 1. An example of the application of the proposed method in the field of cultural heritage digitalization, (a) point cloud; (b) mesh.
Figure 1. An example of the application of the proposed method in the field of cultural heritage digitalization, (a) point cloud; (b) mesh.
Sensors 19 01278 g001
Figure 2. Pipeline of the proposed method in 2D. From left to right are the input point cloud, Delaunay tetrahedra, the s-t graph, the energy minimization result and the final surface mesh.
Figure 2. Pipeline of the proposed method in 2D. From left to right are the input point cloud, Delaunay tetrahedra, the s-t graph, the energy minimization result and the final surface mesh.
Sensors 19 01278 g002
Figure 3. Typical visibility model and soft visibility model in 2D. (a) Typical visibility model; and (b) soft visibility model for a single line of sight v; how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) or its extension are shown.
Figure 3. Typical visibility model and soft visibility model in 2D. (a) Typical visibility model; and (b) soft visibility model for a single line of sight v; how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) or its extension are shown.
Sensors 19 01278 g003
Figure 4. Our visibility model and end tetrahedra comparison in 2D. (a) In our visibility model, for a single line of sight v, how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) is shown. (b) From left to right: Typical end tetrahedron of noise points and that of true surface points on densely sampled surfaces.
Figure 4. Our visibility model and end tetrahedra comparison in 2D. (a) In our visibility model, for a single line of sight v, how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) is shown. (b) From left to right: Typical end tetrahedron of noise points and that of true surface points on densely sampled surfaces.
Sensors 19 01278 g004
Figure 5. Surface reconstruction without and with the likelihood energy. From left to right: (a) point cloud with heavy noise; (b) reconstructed meshes without; and (c) with the likelihood energy.
Figure 5. Surface reconstruction without and with the likelihood energy. From left to right: (a) point cloud with heavy noise; (b) reconstructed meshes without; and (c) with the likelihood energy.
Sensors 19 01278 g005
Figure 6. Free-space support analysis. The four graphs show the ratios of number of outside and inside tetrahedra in different percentiles of all free-space support scores. The four datasets are: (a) Fountain-P11 [3]; (b) Herz-Jesu-P8 [3]; (c) Temple-P312 [1]; and (d) scan23-P49 [2].
Figure 6. Free-space support analysis. The four graphs show the ratios of number of outside and inside tetrahedra in different percentiles of all free-space support scores. The four datasets are: (a) Fountain-P11 [3]; (b) Herz-Jesu-P8 [3]; (c) Temple-P312 [1]; and (d) scan23-P49 [2].
Sensors 19 01278 g006
Figure 7. Free-space support threshold evaluation. The true positive rates and false positive rates in the same four datasets as in Figure 6 are evaluated by setting different free-space support thresholds.
Figure 7. Free-space support threshold evaluation. The true positive rates and false positive rates in the same four datasets as in Figure 6 are evaluated by setting different free-space support thresholds.
Sensors 19 01278 g007
Figure 8. Object edges in the reconstructed meshes, the original image and the depth map. From left to right: (a) the object edges in surface meshes reconstructed by the method in [5] (top) and our method (bottom) without the dense visibility technique; (b) the original image and the corresponding depth map with a similar view point; and (c) the object edges in surface meshes reconstructed by the method in [5] (top) and our method (bottom) with the dense visibility technique.
Figure 8. Object edges in the reconstructed meshes, the original image and the depth map. From left to right: (a) the object edges in surface meshes reconstructed by the method in [5] (top) and our method (bottom) without the dense visibility technique; (b) the original image and the corresponding depth map with a similar view point; and (c) the object edges in surface meshes reconstructed by the method in [5] (top) and our method (bottom) with the dense visibility technique.
Sensors 19 01278 g008
Figure 9. The modified version of our visibility model in 2D. In this visibility model, for a single line of sight v, how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) or its extension is shown.
Figure 9. The modified version of our visibility model in 2D. In this visibility model, for a single line of sight v, how to assign weight (energy) to the tetrahedron containing camera center, the end tetrahedron and the facets intersected by ( c , p ) or its extension is shown.
Sensors 19 01278 g009
Figure 10. Result of the five methods on MVS dataset [2]. GT is the reference model; Tol is Tola et al. [27]; Fur is Furukawa and Ponce [29]; Cam is Campbell et al. [24]; Jan is Jancosek and Pajdla [5]; Our is the proposed method; and ∗_Sur is the surface generated by poisson surface reconstruction method [30].
Figure 10. Result of the five methods on MVS dataset [2]. GT is the reference model; Tol is Tola et al. [27]; Fur is Furukawa and Ponce [29]; Cam is Campbell et al. [24]; Jan is Jancosek and Pajdla [5]; Our is the proposed method; and ∗_Sur is the surface generated by poisson surface reconstruction method [30].
Sensors 19 01278 g010
Figure 11. Quantitative evaluation (lower is better) of the methods on MVS dataset [2]. The abbreviations of the methods are given in Figure 10. In addition, OMVS is OpenMVS; Our∗ is the proposed method without the dense visibility technique; and ∗_Pts is the point cloud generated by method ∗.
Figure 11. Quantitative evaluation (lower is better) of the methods on MVS dataset [2]. The abbreviations of the methods are given in Figure 10. In addition, OMVS is OpenMVS; Our∗ is the proposed method without the dense visibility technique; and ∗_Pts is the point cloud generated by method ∗.
Sensors 19 01278 g011
Figure 12. Detailed views of three methods on MVS dataset [2]. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method. In the even rows are the evaluation result (lower is better) of the corresponding local models in the odd rows through the method in [41]. The unit is mm for all numbers.
Figure 12. Detailed views of three methods on MVS dataset [2]. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method. In the even rows are the evaluation result (lower is better) of the corresponding local models in the odd rows through the method in [41]. The unit is mm for all numbers.
Sensors 19 01278 g012
Figure 13. Result of the proposed method on four scenes of the training set of Tanks and Temples dataset [40]. From left to right are: (a) the input images; (b) the precision of the model generated by the proposed method; (c) the recall of the model generated by the proposed method; and (d) the evaluation result (higher is better) of the models generated by Colmap [38] and three other methods depicted in Figure 10.
Figure 13. Result of the proposed method on four scenes of the training set of Tanks and Temples dataset [40]. From left to right are: (a) the input images; (b) the precision of the model generated by the proposed method; (c) the recall of the model generated by the proposed method; and (d) the evaluation result (higher is better) of the models generated by Colmap [38] and three other methods depicted in Figure 10.
Sensors 19 01278 g013
Figure 14. Detailed views of three methods on four scenes of the training set of Tanks and Temples dataset [40]. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method. In the even rows are the evaluation result (lower is better) of the corresponding local models in the odd rows through the method in [41]. The unit is m for all numbers.
Figure 14. Detailed views of three methods on four scenes of the training set of Tanks and Temples dataset [40]. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method. In the even rows are the evaluation result (lower is better) of the corresponding local models in the odd rows through the method in [41]. The unit is m for all numbers.
Sensors 19 01278 g014
Figure 15. Detailed views of three methods on the intermediate set of Tanks and Temples dataset [40]. From top to bottom are Family, Francis, Horse, Lighthouse, M60, Panther, Playground and Train. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method.
Figure 15. Detailed views of three methods on the intermediate set of Tanks and Temples dataset [40]. From top to bottom are Family, Francis, Horse, Lighthouse, M60, Panther, Playground and Train. From left to right are: (a) the point cloud generated by OpenMVS; (b) the mesh of Jancosek and Pajdla [5]; (c) the mesh of the proposed method without the dense visibility technique; and (d) the mesh of the proposed method.
Sensors 19 01278 g015
Table 1. Symbols used in this work.
Table 1. Symbols used in this work.
SymbolMeaning
vline of sight
ccamera center (a 3D point)
p3D point
Ttetrahedron
l T label of tetrahedron T
D ( l T ) unary energy of the label assignment of tetrahedron T
W ( l T i , l T j ) pair-wise energy of the label assignments of two adjacent tetrahedra
α v weight of a line of sight v
N v amount of tetrahedra intersected with a line of sight v
ddistance between point p and the intersecting point of a segment and a facet
σ scale factor
rthe radius of the circumsphere of the end tetrahedron
U o u t ( T ) energy of tetrahedron T being labeled as outside
U i n ( T ) energy of tetrahedron T being labeled as inside
f ( T ) free-space support of tetrahedron T
β constant for transferring f ( T )
λ , λ v i s , λ l i k e , λ q u a l balance factor
E , E v i s , E l i k e , E q u a l , E v i s t y p i c a l energy
w f weight of a facet f
ϕ , ψ angle
Table 2. Information of the 3D models evaluated on MVS dataset [2]. ∗_Pts is the number of points in a 3D point cloud. ∗_Vtx and ∗_Fcs are the number of vertices and facets of a 3D mesh, respectively. The unit of all numbers is million.
Table 2. Information of the 3D models evaluated on MVS dataset [2]. ∗_Pts is the number of points in a 3D point cloud. ∗_Vtx and ∗_Fcs are the number of vertices and facets of a 3D mesh, respectively. The unit of all numbers is million.
SceneID1234569101521232429364461110114118122
Tol_Pts1.01.10.90.70.91.01.00.71.01.01.10.80.71.10.90.70.71.21.00.9
Tol_Vtx2.12.22.11.81.92.32.61.72.32.12.32.41.12.01.51.61.62.12.11.8
Tol_Fcs4.24.44.23.53.74.65.23.34.74.34.54.82.14.13.13.23.24.24.23.5
Fur_Pts2.32.62.52.22.22.42.41.92.53.03.12.52.32.72.71.62.22.62.62.4
Fur_Vtx1.11.11.20.80.80.71.00.72.52.92.81.02.11.91.70.91.81.51.71.4
Fur_Fcs2.22.22.41.61.61.52.01.54.95.85.51.94.23.83.41.73.62.93.32.7
Cam_Pts23.629.622.220.820.223.619.813.022.024.029.520.216.529.520.27.619.926.130.221.7
Cam_Vtx4.24.68.14.86.86.716.02.612.08.94.12.63.33.25.13.76.35.131.26.1
Cam_Fcs8.59.216.39.513.513.432.05.124.017.88.25.26.66.310.27.312.510.262.412.1
OMVS_Pts11.811.012.210.111.810.99.18.39.210.112.29.07.811.39.88.98.013.18.88.4
Jan_Vtx0.60.60.70.50.60.60.60.50.60.80.70.60.60.70.70.50.50.70.60.6
Jan_Fcs1.31.21.41.11.21.21.21.01.31.61.41.21.21.51.50.91.01.31.21.1
Our*_Vtx1.21.11.11.00.91.01.00.91.11.31.31.21.01.31.10.71.01.10.91.0
Our*_Fcs2.52.32.22.01.82.02.01.92.22.62.62.42.02.62.31.51.92.21.81.9
Our_Vtx1.61.61.41.01.21.41.31.21.51.51.71.31.31.31.20.71.01.51.21.1
Our_Fcs3.33.22.92.02.52.92.62.43.03.13.42.62.72.62.51.42.03.12.42.3
Table 3. Evaluation result of the 3D models evaluated on the training set of Tanks and Temples dataset [40] through the method in [42]. M stands for million. Precision is expressed as a proportion 1:k, where k is the size of the scene divided by the standard error. The unit of all other numbers is mm.
Table 3. Evaluation result of the 3D models evaluated on the training set of Tanks and Temples dataset [40] through the method in [42]. M stands for million. Precision is expressed as a proportion 1:k, where k is the size of the scene divided by the standard error. The unit of all other numbers is mm.
TypeBarnIgnatius
ColmapOMVS_PtsJanOur*OurColmapOMVS_PtsJanOur*Our
Pts6.2M35.4M6.3M5.8M6.7M1.3M13.1M3.5M2.9M3.3M
mean19.2417.7010.4411.1410.232.663.552.512.902.14
95.5%<59.9334.7528.1729.4226.047.9612.156.337.695.15
99.7%<221.83181.46118.49120.25113.4539.1835.9138.5538.4934.38
Precision1:8001:10001:14001:14001:15001:5001:6001:8001:8001:900
TypeCourthouseTruck
ColmapOMVS_PtsJanOur*OurColmapOMVS_PtsJanOur*Our
Pts17.3M63.4M14.4M13.7M15.0M3.8M22.5M3.7M3.0M3.5M
mean97.56247.9493.1596.8491.888.077.296.657.136.47
95.5%<315.51597.98302.36311.29294.3124.6723.6221.4823.8721.46
99.7%<2261.965108.952086.732117.382029.45154.56174.76147.83151.72143.58
Precision1:3001:2001:4001:4001:4001:4001:4001:6001:6001:700
Table 4. Leaderboard  1 of the methods and the result of the proposed method with respect to F-score on the intermediate set of Tanks and Temples dataset [40].
Table 4. Leaderboard  1 of the methods and the result of the proposed method with respect to F-score on the intermediate set of Tanks and Temples dataset [40].
MethodFamilyFrancisHorseLighthouseM60PantherPlaygroundTrainMean
PMVSNet70.0444.6440.2265.2055.0855.1760.3754.2955.62
Altizure-HKUST74.6061.3038.4861.4854.9353.3256.2149.4756.22
ACMH69.9949.4545.1259.0452.6452.3758.3451.6154.82
Dense R-MVSNet73.0154.4643.4243.8846.8046.6950.8745.2550.55
R-MVSNet69.9646.6532.5942.9551.8848.8052.0042.3848.40
i23dMVS456.6433.7528.4048.4239.2344.8748.3437.8842.19
MVSNet55.9928.5525.0750.7953.9650.8647.9034.6943.48
COLMAP50.4122.2525.6356.4344.8346.9748.5342.0442.14
Pix4D64.4531.9126.4354.4150.5835.3747.7834.9643.24
i23dMVS_356.2133.1428.9247.7440.2944.2046.9337.6641.89
OpenMVG + OpenMVS58.8632.5926.2543.1244.7346.8545.9735.2741.71
OpenMVG + MVE49.9128.1920.7543.3544.5144.7636.5835.9538.00
OpenMVG + SMVS31.9319.9215.0239.3836.5141.6135.8925.1230.67
Theia-I + OpenMVS48.1119.3820.6630.0230.3730.7923.6520.4627.93
OpenMVG + PMVS41.0317.7012.8336.6835.9333.2031.7828.1029.66
Jan62.6947.4434.5257.9438.6747.0655.2639.9047.94
Our*62.4646.6832.6157.6633.6644.2552.4038.2546.00
Our65.2149.4135.4159.0437.5747.8556.7741.2849.07
Table 5. Information of the 3D models evaluated on the intermediate set of Tanks and Temples dataset [40]. ∗_Pts is the number of points in a 3D point cloud. ∗_Vtx and ∗_Fcs are the number of vertices and facets of a 3D mesh, respectively. The unit of all numbers is million.
Table 5. Information of the 3D models evaluated on the intermediate set of Tanks and Temples dataset [40]. ∗_Pts is the number of points in a 3D point cloud. ∗_Vtx and ∗_Fcs are the number of vertices and facets of a 3D mesh, respectively. The unit of all numbers is million.
SceneFamilyFrancisHorseLighthouseM60PantherPlaygroundTrain
OMVS_Pts12.017.89.028.424.826.028.631.9
Jan_Vtx2.01.91.52.85.14.15.74.8
Jan_Fcs4.13.73.05.610.18.311.39.6
Our∗_Vtx1.61.21.21.73.62.94.13.2
Our∗_Fcs3.22.42.33.57.25.88.26.5
Our_Vtx2.52.72.13.35.35.46.96.3
Our_Fcs5.15.44.26.610.610.813.912.5

Share and Cite

MDPI and ACS Style

Zhou, Y.; Shen, S.; Hu, Z. Detail Preserved Surface Reconstruction from Point Cloud. Sensors 2019, 19, 1278. https://doi.org/10.3390/s19061278

AMA Style

Zhou Y, Shen S, Hu Z. Detail Preserved Surface Reconstruction from Point Cloud. Sensors. 2019; 19(6):1278. https://doi.org/10.3390/s19061278

Chicago/Turabian Style

Zhou, Yang, Shuhan Shen, and Zhanyi Hu. 2019. "Detail Preserved Surface Reconstruction from Point Cloud" Sensors 19, no. 6: 1278. https://doi.org/10.3390/s19061278

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop