scholarly journals Automatic Segmentation for Plant Leaves via Multiview Stereo Reconstruction

2017 ◽  
Vol 2017 ◽  
pp. 1-11 ◽  
Author(s):  
Jingwei Guo ◽  
Lihong Xu

This paper presented a new method for automatic plant point cloud acquisition and leaves segmentation. Quasi-dense point cloud of the plant is obtained from multiview stereo reconstruction based on surface expansion. In order to overcome the negative effects from complex natural light changes and to obtain a more accurate plant point cloud, the Adaptive Normalized Cross-Correlation algorithm is used in calculating the matching cost between two images, which is robust to radiometric factors and can reduce the fattening effect around boundaries. In the stage of segmentation for each single leaf, an improved region growing method based on fully connected conditional random field (CRF) is proposed to separate the connected leaves with similar color. The method has three steps: boundary erosion, initial segmentation, and segmentation refinement. First, the edge of each leaf point cloud is eroded to remove the connectivity between leaves. Then leaves will be initially segmented by region growing algorithm based on local surface normal and curvature. Finally an efficient CRF inference method based on mean field approximation is applied to remove small isolated regions. Experimental results show that our multiview stereo reconstruction method is robust to illumination changes and can obtain accurate color point clouds. And the improved region growing method based on CRF can effectively separate the connected leaves in obtained point cloud.

Author(s):  
M. Bassier ◽  
M. Bonduel ◽  
B. Van Genechten ◽  
M. Vergauwen

Point cloud segmentation is a crucial step in scene understanding and interpretation. The goal is to decompose the initial data into sets of workable clusters with similar properties. Additionally, it is a key aspect in the automated procedure from point cloud data to BIM. Current approaches typically only segment a single type of primitive such as planes or cylinders. Also, current algorithms suffer from oversegmenting the data and are often sensor or scene dependent.<br><br> In this work, a method is presented to automatically segment large unstructured point clouds of buildings. More specifically, the segmentation is formulated as a graph optimisation problem. First, the data is oversegmented with a greedy octree-based region growing method. The growing is conditioned on the segmentation of planes as well as smooth surfaces. Next, the candidate clusters are represented by a Conditional Random Field after which the most likely configuration of candidate clusters is computed given a set of local and contextual features. The experiments prove that the used method is a fast and reliable framework for unstructured point cloud segmentation. Processing speeds up to 40,000 points per second are recorded for the region growing. Additionally, the recall and precision of the graph clustering is approximately 80%. Overall, nearly 22% of oversegmentation is reduced by clustering the data. These clusters will be classified and used as a basis for the reconstruction of BIM models.


2019 ◽  
Vol 11 (23) ◽  
pp. 2727 ◽  
Author(s):  
Ming Huang ◽  
Pengcheng Wei ◽  
Xianglei Liu

Plane segmentation is a basic yet important process in light detection and ranging (LiDAR) point cloud processing. The traditional point cloud plane segmentation algorithm is typically affected by the number of point clouds and the noise data, which results in slow segmentation efficiency and poor segmentation effect. Hence, an efficient encoding voxel-based segmentation (EVBS) algorithm based on a fast adjacent voxel search is proposed in this study. First, a binary octree algorithm is proposed to construct the voxel as the segmentation object and code the voxel, which can compute voxel features quickly and accurately. Second, a voxel-based region growing algorithm is proposed to cluster the corresponding voxel to perform the initial point cloud segmentation, which can improve the rationality of seed selection. Finally, a refining point method is proposed to solve the problem of under-segmentation in unlabeled voxels by judging the relationship between the points and the segmented plane. Experimental results demonstrate that the proposed algorithm is better than the traditional algorithm in terms of computation time, extraction accuracy, and recall rate.


2019 ◽  
Vol 11 (13) ◽  
pp. 1586 ◽  
Author(s):  
Maarten Bassier ◽  
Maarten Vergauwen

The automated reconstruction of Building Information Modeling (BIM) objects from point cloud data is still subject of ongoing research. A vital step in the process is identifying the observations for each wall object. Given a set of segmented and classified point clouds, the labeled segments should be clustered according to their respective objects. The current processes to perform this task are sensitive to noise, occlusions, and the associativity between faces of neighboring objects. The proper retrieval of the observed geometry is especially important for wall geometry as it forms the basis for further BIM reconstruction. In this work, a method is presented to automatically group wall segments derived from point clouds according to the proper walls of a building. More specifically, a Conditional Random Field is employed that evaluates the context of each wall segment in order to determine which wall it belongs to. First, a set of classified planar primitives is obtained through algorithms developed in prior work. Next, both local and contextual features are extracted based on the nearest neighbors and a number of seeds that are heuristically determined. The final wall clusters are then computed by decoding the graph. The method is tested on our own data as well as the 2D-3D-Semantics (2D-3D-S) benchmark data of Stanford. Compared to a conventional region growing method, the proposed method reduces the rate of false positives, resulting in better wall clusters. Overall, the method computes a more balanced clustering of the observations. A key advantage of the proposed method is its capability to deal with wall geometry in complex configurations in multi-storey buildings opposed to the presented methods in current literature.


Sensors ◽  
2018 ◽  
Vol 18 (11) ◽  
pp. 3625 ◽  
Author(s):  
Dawei Li ◽  
Yan Cao ◽  
Xue-song Tang ◽  
Siyuan Yan ◽  
Xin Cai

Leaves account for the largest proportion of all organ areas for most kinds of plants, and are comprise the main part of the photosynthetically active material in a plant. Observation of individual leaves can help to recognize their growth status and measure complex phenotypic traits. Current image-based leaf segmentation methods have problems due to highly restricted species and vulnerability toward canopy occlusion. In this work, we propose an individual leaf segmentation approach for dense plant point clouds using facet over-segmentation and facet region growing. The approach can be divided into three steps: (1) point cloud pre-processing, (2) facet over-segmentation, and (3) facet region growing for individual leaf segmentation. The experimental results show that the proposed method is effective and efficient in segmenting individual leaves from 3D point clouds of greenhouse ornamentals such as Epipremnum aureum, Monstera deliciosa, and Calathea makoyana, and the average precision and recall are both above 90%. The results also reveal the wide applicability of the proposed methodology for point clouds scanned from different kinds of 3D imaging systems, such as stereo vision and Kinect v2. Moreover, our method is potentially applicable in a broad range of applications that aim at segmenting regular surfaces and objects from a point cloud.


Author(s):  
A. Kharroubi ◽  
R. Hajji ◽  
R. Billen ◽  
F. Poux

Abstract. With the increasing volume of 3D applications using immersive technologies such as virtual, augmented and mixed reality, it is very interesting to create better ways to integrate unstructured 3D data such as point clouds as a source of data. Indeed, this can lead to an efficient workflow from 3D capture to 3D immersive environment creation without the need to derive 3D model, and lengthy optimization pipelines. In this paper, the main focus is on the direct classification and integration of massive 3D point clouds in a virtual reality (VR) environment. The emphasis is put on leveraging open-source frameworks for an easy replication of the findings. First, we develop a semi-automatic segmentation approach to provide semantic descriptors (mainly classes) to groups of points. We then build an octree data structure leveraged through out-of-core algorithms to load in real time and continuously only the points that are in the VR user's field of view. Then, we provide an open-source solution using Unity with a user interface for VR point cloud interaction and visualisation. Finally, we provide a full semantic VR data integration enhanced through developed shaders for future spatio-semantic queries. We tested our approach on several datasets of which a point cloud composed of 2.3 billion points, representing the heritage site of the castle of Jehay (Belgium). The results underline the efficiency and performance of the solution for visualizing classifieds massive point clouds in virtual environments with more than 100 frame per second.


Author(s):  
M. Bassier ◽  
M. Vergauwen

<p><strong>Abstract.</strong> The automated reconstruction of Building Information Modeling (BIM) objects from point cloud data is still ongoing research. A key aspect is retrieving the proper observations for each object. After segmenting and classifying the initial point cloud, the labeled segments should be clustered according to their respective objects. However, this procedure is challenging due to noise, occlusions and the associativity between different objects. This is especially important for wall geometry as it forms the basis for further BIM reconstruction.</p><p> In this work, a method is presented to automatically group wall segments derived from point clouds according to the proper walls of a building. More specifically, a Conditional Random Field is employed that evaluates the context of each observation in order to determine which wall it belongs too. The emphasis is on the clustering of highly associative walls as this topic currently is a gap in the body of knowledge. First a set of classified planar primitives is obtained using algorithms developed in prior work. Next, both local and contextual features are extracted based on the nearest neighbors and a number of seeds that are heuristically determined. The final wall clusters are then computed by decoding the graph and thus the most likely configuration of the observations. The experiments prove that the used method is a promising framework for wall clustering from unstructured point cloud data. Compared to a conventional region growing method, the proposed method significantly reduces the rate of false positives, resulting in better wall clusters. A key advantage of the proposed method is its capability of dealing with complex wall geometry in entire buildings opposed to the presented methods in current literature.</p>


2021 ◽  
Vol 14 (1) ◽  
pp. 95
Author(s):  
Zhonghua Su ◽  
Zhenji Gao ◽  
Guiyun Zhou ◽  
Shihua Li ◽  
Lihui Song ◽  
...  

Planes are essential features to describe the shapes of buildings. The segmentation of a plane is significant when reconstructing a building in three dimensions. However, there is a concern about the accuracy in segmenting plane from point cloud data. The objective of this paper was to develop an effective segmentation algorithm for building planes that combines the region growing algorithm with the distance algorithm based on boundary points. The method was tested on point cloud data from a cottage and pantry as scanned using a Faro Focus 3D laser range scanner and Matterport Camera, respectively. A coarse extraction of the building plane was obtained from the region growing algorithm. The coplanar points where two planes intersect were obtained from the distance algorithm. The building plane’s optimal segmentation was then obtained by combining the coarse extraction plane points and the corresponding coplanar points. The results show that the proposed method successfully segmented the plane points of the cottage and pantry. The optimal distance thresholds using the proposed method from the uncoarse extraction plane points to each plane boundary point of cottage and pantry were 0.025 m and 0.030 m, respectively. The highest correct rate and the highest error rate of the cottage’s (pantry’s) plane segmentations using the proposed method under the optimal distance threshold were 99.93% and 2.30% (98.55% and 2.44%), respectively. The F1 score value of the cottage’s and pantry’s plane segmentations using the proposed method under the optimal distance threshold reached 97.56% and 95.75%, respectively. This method can segment different objects on the same plane, while the random sample consensus (RANSAC) algorithm causes the plane to become over-segmented. The proposed method can also extract the coplanar points at the intersection of two planes, which cannot be separated using the region growing algorithm. Although the RANSAC-RG method combining the RANSAC algorithm and the region growing algorithm can optimize the segmentation results of the RANSAC (region growing) algorithm and has little difference in segmentation effect (especially for cottage data) with the proposed method, the method still loses coplanar points at some intersection of the two planes.


Author(s):  
J. Niemeyer ◽  
F. Rottensteiner ◽  
U. Soergel ◽  
C. Heipke

We propose a novel hierarchical approach for the classification of airborne 3D lidar points. Spatial and semantic context is incorporated via a two-layer Conditional Random Field (CRF). The first layer operates on a point level and utilises higher order cliques. Segments are generated from the labelling obtained in this way. They are the entities of the second layer, which incorporates larger scale context. The classification result of the segments is introduced as an energy term for the next iteration of the point-based layer. This framework iterates and mutually propagates context to improve the classification results. Potentially wrong decisions can be revised at later stages. The output is a labelled point cloud as well as segments roughly corresponding to object instances. Moreover, we present two new contextual features for the segment classification: the &lt;i&gt;distance&lt;/i&gt; and the &lt;i&gt;orientation of a segment with respect to the closest road&lt;/i&gt;. It is shown that the classification benefits from these features. In our experiments the hierarchical framework improve the overall accuracies by 2.3% on a point-based level and by 3.0% on a segment-based level, respectively, compared to a purely point-based classification.


Sensors ◽  
2020 ◽  
Vol 20 (15) ◽  
pp. 4209
Author(s):  
Dongbo Yu ◽  
Jun Xiao ◽  
Ying Wang

In respect of rock-mass engineering, the detection of planar structures from the rock-mass point clouds plays a crucial role in the construction of a lightweight numerical model, while the establishment of high-quality models relies on the accurate results of surface analysis. However, the existing techniques are barely capable to segment the rock mass thoroughly, which is attributed to the cluttered and unpredictable surface structures of the rock mass. This paper proposes a high-precision plane detection approach for 3D rock-mass point clouds, which is effective in dealing with the complex surface structures, thus achieving a high level of detail in detection. Firstly, the input point cloud is fast segmented to voxels using spatial grids, while the local coplanarity test and the edge information calculation are performed to extract the major segments of planes. Secondly, to preserve as much detail as possible, supervoxel segmentation instead of traditional region growing is conducted to deal with scattered points. Finally, a patch-based region growing strategy applicable to rock mass is developed, while the completed planes are obtained by merging supervoxel patches. In this paper, an artificial icosahedron point cloud and four rock-mass point clouds are applied to validate the performance of the proposed method. As indicated by the experimental results, the proposed method can make high-precision plane detection achievable for rock-mass point clouds while ensuring high recall rate. Furthermore, the results of both qualitative and quantitative analyses evidence the superior performance of our algorithm.


Sign in / Sign up

Export Citation Format

Share Document