Next Article in Journal
Multi-Zone Authentication and Privacy-Preserving Protocol (MAPP) Based on the Bilinear Pairing Cryptography for 5G-V2X
Next Article in Special Issue
An Evaluation of Different NIR-Spectral Pre-Treatments to Derive the Soil Parameters C and N of a Humus-Clay-Rich Soil
Previous Article in Journal
A Novel Hybrid Approach for Risk Evaluation of Vehicle Failure Modes
Previous Article in Special Issue
Weed and Corn Seedling Detection in Field Based on Multi Feature Fusion and Support Vector Machine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of 3D Point Clouds of Oilseed Rape Plants Based on Time-of-Flight Cameras

1
College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China
2
Key Laboratory of Spectroscopy Sensing, Ministry of Agriculture and Rural Affairs, Hangzhou 310058, China
3
State Key Laboratory of Modern Optical Instrumentation, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(2), 664; https://doi.org/10.3390/s21020664
Submission received: 17 October 2020 / Revised: 7 January 2021 / Accepted: 16 January 2021 / Published: 19 January 2021
(This article belongs to the Special Issue Sensing Technologies for Agricultural Automation and Robotics)

Abstract

:
Three-dimensional (3D) structure is an important morphological trait of plants for describing their growth and biotic/abiotic stress responses. Various methods have been developed for obtaining 3D plant data, but the data quality and equipment costs are the main factors limiting their development. Here, we propose a method to improve the quality of 3D plant data using the time-of-flight (TOF) camera Kinect V2. A K-dimension (k-d) tree was applied to spatial topological relationships for searching points. Background noise points were then removed with a minimum oriented bounding box (MOBB) with a pass-through filter, while outliers and flying pixel points were removed based on viewpoints and surface normals. After being smoothed with the bilateral filter, the 3D plant data were registered and meshed. We adjusted the mesh patches to eliminate layered points. The results showed that the patches were closer. The average distance between the patches was 1.88 × 10−3 m, and the average angle was 17.64°, which were 54.97% and 48.33% of those values before optimization. The proposed method performed better in reducing noise and the local layered-points phenomenon, and it could help to more accurately determine 3D structure parameters from point clouds and mesh models.

1. Introduction

With the increasing demand for accelerating plant breeding and improving crop-management efficiency, it is necessary to measure various phenotypic traits of plants in a high-throughput and accurate manner [1]. The fast development of advanced sensors and automation and computation tools further promotes the capability and throughput of plant-phenotyping techniques, which allows the nondestructive measurement of complex plant parameters or traits [2]. Plant three-dimensional (3D) morphological structure is an important descriptive trait of plant growth and development, as well as biotic/abiotic stress responses [3]. 3D plant phenotyping has great potential for multiscale analyses of the 3D morphological structures of plant organs, individuals and canopies; for building functional structure plant models (FSPM) [4], for evaluating the performance of different genotypes in adaptation to the environment, for predicting yield potential [5] and for facilitating the accurate management of breeding and crop production with key technical support.
Different 3D sensors and imaging techniques have been developed to quantify plants’ 3D morphological structural parameters at different scales. These sensors can be classified into passive and active sensors [6]. Generally, passive sensors build a 3D model from the images of different views. Some systems have been developed for obtaining a 3D model, such as an RGB camera combined with a structure from motion (SFM) algorithm and a multiview stereo vision system [7,8]. Rose et al. [9] found that the SFM-based photogrammetric method can yield high correlations to the measurements and was suitable for the task of organ-level plant phenotyping. Xiang et al. [10] developed a PhenoStereo system for field-based plant phenotyping and used a set of customized strobe lights for lighting influence. Rossi et al. [11] provided references for optimizing the reconstruction process of SFM in terms of input and time requirements. They found the proper balance between number of images and their quality for an efficient and accurate measurement of individual structural parameters for species with different canopy structures. However, methods combined with passive sensors have high requirements for images with complex features in the surface texture for image matching [6], and the methods are limited by lighting condition as well as the complexity of algorithm.
Active sensors acquire distance information from the active emission of signals [12]. Laser scanning is considered to be a universal, high-precision and wide-scale detection method for plant-growth status [5]. Paulus et al. [13] conducted a growth analysis experiment on eight pots of spring barley plants under different drought conditions in an industrial environment. Single leaf area, single stem height, plant height and plant width were determined with a laser-scanning system combined with an articulated measuring arm. These measurements had high correlations (R2, 0.85–0.97) with manual measurements. Based on such accuracy, they were also able to effectively monitor the growth and quantify the growth processes of barley plants. However, the small scanning field and small arm size necessitated multiple-location scans for whole plants, which made the system expensive and inefficient. Sun et al. [14] developed a system consisting of a 2D light detection and ranging (LiDAR) and real time kinematic global positioning system (RTK-GPS) for high-throughput phenotyping. They built a model to obtain the height of cotton plants, considering the angular resolution, sensor mounting height, tractor speed and so on. This system performed well in estimating the heights of cotton plants. However, many factors such as the angular resolution and uneven ground affected the measurements, and the data were noisy, which made it impossible to accurately measure other parameters such as the leaf area. Su et al. [15] proposed a difference-of-normals (DoN) method to separate corn leaves and stalks based on laser point clouds in a greenhouse. However, it took 20 min for each scan per position. Ana et al. [16] proposed a vine-shaped artificial object (VSAO) calibration method, based on which they implemented a static terrestrial laser scanner (TLS) and a mobile scanning system (MMS) with six algorithms to determine the trunk volumes of vines in a real vineyard. The results showed that the relative errors of the different sensors, combined with different algorithms, were 2.46%–6.40%. The limitations of these two systems included long scanning time, tedious processing and environmental factors. Laser scanner had high detection accuracy for individual plants and groups in industrial or field environments, but many factors such as the topography still impacted the measurements [17]. However, cost and the efficiency were the mainly bottlenecks that restricted the application of this technology in actual production.
Some other detection methods have been proposed, and laser scanning has become a common means of evaluating these methods [18,19,20]. Compared to laser scanning, the time-of-flight (TOF) camera has the advantages of speed, simplicity and low cost and the potential for use in 3D phenotyping research [20,21,22,23,24,25,26]. For example, Microsoft Kinect is widely used as a typical TOF camera. Paulus et al. [19] proved that a low-cost system based on the Microsoft Kinect device can effectively estimate the phenotypes of sugar beets. They used the David laser scanner system as a reference method. The results showed that the Kinect performed as well as the laser scanner for sugar-beet taproots in terms of height, width, volume and surface area estimation. However, the Kinect performed poorly in estimating wheat-ear parameters, due to the low resolution, while the laser scanner still performed well. The R2s of the maximum length and alpha shape volume were 0.02 and 0.40, respectively, when using Kinect, and the R2s of these two parameters were greater than 0.84 for the laser scan. Sugar beet is simple in morphology and structure; the potential of Kinect for other plants remains to be seen. Xia et al. [27] used a mean-shift-clustering algorithm to segment the leaves in depth images obtained from Kinect, and removed the background in both RGB and depth images. Based on the adjacent-pixel-gradient-vector-field of depth image, they achieved segmentation of shade leaves. This approach can be effectively applied to automatic fruit harvest and other agricultural automation work. However, their work only focused on a single-frame point cloud, which led to incomplete data for the plant. Meanwhile, the complete plant point clouds were more complex, with more noise and a layered-points phenomenon, which their algorithm could not solve. Anduja et al. [28] proposed reconstructing maize in the field with the Kinect Fusion algorithm. They monitored segmentation of maize, weeds and soil through height and RGB information and studied the correlation between volume and biomass. The coefficient of the correlation between the maize biomass and volume was 0.77, while that between the weed volume and biomass was 0.83. It was clear that the correlation coefficient was not as high as that in Paulus’s study [19] because of rough point clouds with poor quality caused by the complex field environment and complexity of the plants, and they did not perform point-cloud optimization. Wang et al. [29] measured the height of sorghum in the field using five different sensors and established digital elevation models. All the coefficients of correlation between the values generated by the models and those measured manually were above 0.9. They proposed that the Kinect could provide color and morphology information about plants for identification and counting. However, the data acquired by Kinect were, again, rough and noisy, and they were not suitable for the extraction of other parameters.
According to the above studies, multiple complex parameters were effectively extracted using laser scans because of the high-quality point clouds. Kinect, by contrast, performed well in height estimation and object segmentation because these two tasks do not require high-quality data. To extract more parameters efficiently in a low-cost platform, it was necessary to obtain complete and high-quality plant 3D data using a TOF camera. However, there was a layered-points phenomenon in the plant point clouds based on multiple frames [30] because of the errors from the TOF camera and registration algorithm, a common problem.
To improve the quality of the plant point cloud, we proposed an optimization method to reduce the impact of noise and layered-points. A simple and low-cost platform based on Kinect was used for data acquisition, which makes the proposed method more widely applicable. In this study, we optimized the quality of single-frame point clouds by removing all types of noise while preserving the integrity of the plant data. We also eliminated the local layered-points phenomenon to improve the quality of plant point clouds registered from multiple frames.

2. Materials and Methods

2.1. Experimental Setup and Data Acquisition

The data used in this study were collected for one oilseed rape cultivar (Brassica napus L. cv. Zhe Da 619) in a closed indoor imaging platform, mainly comprising a Kinect V2 sensor, turntable and computer. The Kinect V2 (Windows version, Microsoft, Redmond, WA, USA) consisted of an RGB camera (1920 × 1080), near-infrared camera (512 × 424), and near-infrared light for acquiring color and depth data, respectively. The acquisition platform and point cloud acquisition are shown in Figure 1. As shown in Figure 1a, the Kinect V2 was about 0.75 m away from the main stem (vertical axis) of the plant, and the shooting angle was 30°. The rotary speed of the turntable was 14.4°/s for changing the plant pose, and the measured plant was placed on the center of the turntable. The computer controlled the Kinect V2 and acquired and processed the raw data. It had an Intel core i5-4590 processor, a Windows 10 64-bit operating system and 8GB of ECC RAM. Data processing was performed on Point Cloud Library (PCL) and Open3D Library in Visual studio 2013 (Professional version, Microsoft, Redmond, WA, USA). Before acquisition, the Kinect V2 camera was calibrated by Zhang’s [31] method, and the transformation matrix between the RGB and depth cameras was adjusted to optimize the mapping relationship for the two types of images to ensure the consistency of the color and depth of each point (Figure 1b).
A point-cloud-processing pipeline was developed to optimize the quality of the entire plant point cloud. As show in Figure 2, the pipeline of workflow mainly comprised three steps: (1) point-cloud noise removal; (2) point-cloud smoothing; (3) registration optimization based on neighboring meshes.

2.1.1. Point-Cloud Noise Removal

The point cloud acquired by Kinect V2 was generally disordered, with many noise points that would have a significant effect on the reconstruction accuracy and computation speed. The viewpoint feature and normal feature of the point cloud were used to remove the noise based on the spatial topological relationship established by the k-dimensional (k-d) tree. The spatial topological relationship was used for searching neighboring points.
There were three types of noise in the point cloud: the background noise (BN), which consisted of nontarget points away from the targets; the outlier noise (ON), which consisted of scattered points, mostly around the targets, caused by the sensors, and the flying pixel noise (FPN) from the boundaries of two objects [32]. Traditionally, the BN has mainly been eliminated with a pass-through filter, and the ON removed based on the neighboring points. The pass-through filter limited the ranges of the X, Y and Z axes and removed the points outside the ranges. Due to FPN, points covering different objects of different depths were distant. The vector made up of FPN and viewer points was almost perpendicular to the FPN point’s normal vector. The FPN points could be removed based on these two features.
Because the central axis of the plant is not strictly perpendicular to the camera-projection direction during data acquisition, it is very difficult to eliminate the BN points while preserving the integrity of the plant by using the traditional pass-through filter method. In this study, a combination of a pass-through filter and minimum oriented bounding box (MOBB) was proposed. The MOBB was a cuboid that contained the object as tightly as possible, with the smallest volume in the defined coordinate system. In 2D space, assuming the camera tilt angle was θ, this coordinate was aligned parallel to the camera coordinate. If the data of the object (red box) had a rectangular distribution under the ideal condition as shown in Figure 3a, the MOBB (black box) was equivalent to this rectangle. In this case, after the MOBB was rotated in the β counterclockwise direction around Point A, the object was aligned and parallel to the camera coordinate, and θ = β . Normally, the distribution of the object was irregular (red box) as shown in Figure 3b. The relationship between the angles was calculated as below:
θ = π 2 α 3 ,
β 2   = β 3 = α 2 + α 3 ;
α 2   = α ;
β 2 = π 2 θ + α ;
where θ is the camera tilt angle, α , α 2 and β are the angles between the MOBB and camera coordinate and α 3 is the angle between the object box and camera coordinate. β 2 and β 3 are the angles between the MOBB and object box and were equivalent in value.
In this case, α and β can be obtained from the MOBB orthogonal and the camera coordinates. After the object was rotated in β 2 counterclockwise around Point A’, firstly, and the MOBB was rotated β counterclockwise around Point A, secondly, the object was aligned and parallel to the camera coordinate. The same method was applied to 3D space, in which the rotation of the object was achieved with Euler’s formula. The MOBB orthogonal coordinate was established with the center of the point cloud data as the coordinate origin and the length, width and height of the MOBB, which can improve the performance of the pass-through filter.
After the removal of the BN points, there were still many ON and FPN points that needed to be removed. A radius-density-based outlier filter was implemented to remove the ON points [33]. For each point p i of data, it takes into account both the number K and the average distance d ¯ ( p ) of the neighboring points within a certain radius r of the selected point. If the selected point was judged as ON, the following conditions were met:
d ¯ ( p ) = 1 K p j N ( p ) p p j
d ¯ ( p ) > ( μ + n · σ )
K > k
where p j is the neighboring point of the selected point p, μ is the average distance between neighboring points, σ is the standard deviation of μ , n is the multiple of σ , and k is the defined point number.
As for FPN, it can be removed based on the angle θ of the normal vector n i and the view vector n v connecting with the viewpoint. The viewer vector consists of the viewpoint n v and p . For each point p , if θ was bigger than the threshold θ a n g l e , this point could be removed as FPN. The noise-elimination process is summarized in Algorithm 1 as shown below.
Algorithm 1: Point-cloud-noise removal
Input: Raw data of point cloud { p i n }
Output: Object point cloud { p f i l t e r o u t } without noise.
(1) Establishing the spatial topological relationship of the source data using the k-d tree.
(2) Obtaining the maximum and minimum values of the three coordinate axes in the point cloud and searching six boundary points with x m i n , y m i n , z m i n , x m a x , y m a x and z m a x respectively. A radius-density-based outlier filter is used on these six points. If they are outliers, then delete and repeat this step. Otherwise, proceed to the next step.
(3) Building up the MOBB, rotating it with Euler’s formula and removing BN points using a pass-through filter.
(4) Removing ON points using the radius-density-based outlier filter for all points.
(5) For each point of data, computing the normal vector n of the selected point p by principal component analysis (PCA). Computing the components of the normal vector n and the view vector n v projection on the x o z plane of the camera coordinate space separately, and then obtaining angle θ through the cosine theorem.
(6) Comparing θ and θ a n g l e and removing FPN points.
(7) Performing Operations 4 through 7 on all points and outputting point { p f i l t e r o u t } .
In order to evaluate the effect of noise removal, the benchmark point cloud was segmented manually in Geomagic Studio [34]. Thus, the valid-point percent (VPP) was proposed. The closer the VPP to 100%, the fewer the non-target points.
V P P = V a l i d   p o i n t s / T o t a l   p o i n t s

2.1.2. Point-Cloud Smoothing

The bilateral filter is a nonlinear filtration tool used for edge-preserving smoothing [35]. Due to the wiggling error caused by the Kinect sensor, the fitting surface of the data acquired by the Kinect was not smooth [22]; this could be solved by this filter. Several 3D bilateral filters are based on the mesh model [36,37]. However, the mesh or fitting surface was easily affected by the noise. Based on the neighboring points, the disordered bilateral filter was used to smooth the point cloud while preserving the edge features of the point cloud [33].
α = p { p i r } ( W c ( p i p ) W s ( p i p , n i ) ( p i p · n i ) ) p { p i r } W c ( p i p ) W s ( p i p , n i )
p = p i + α n i
W c ( p i p ) = e x p [ p i p 2 2 σ c 2 ]
W s ( p i p ) = e x p [ p i p 2 2 σ s 2 ]
where p i is the selected point and { p i r } are the neighboring points within the radius of r . W c is related to the smoothness and σ c is the distance factor. W s is related to the ability to preserve features and σ s is the hue factor.

2.1.3. Registration Optimization Based on Neighboring Meshes

The purpose of registration was to unify the point clouds from different coordinate systems into the same coordinate system [38]. Multiple neighboring point clouds were registered into a single point cloud using fast-point-feature histograms (FPFH) for rough registration and an iterative-closest-point (ICP) algorithm for fine alignment [39]. However, the local layered points could be observed after registration due to the complex refraction and reflection situations in the interiors or surfaces of the leaves [30]. The accuracy of the algorithm also affected the layered-points phenomenon. These stratified leaves were close, and the layered-points phenomenon could be optimized by adjusting the related points’ position. In the point-cloud model, there was no geometrical relationship between the points, and the topological relation supported by the k-d tree was only applicable to searching neighboring points. A mesh model based on a triangular patch was more suitable for solving the issue of layered points. Three definitions were proposed to explain the triangular patch relationship in Figure 4. The symbol stands for a triangle patch.
Definition 1.
Intersecting relationship: there are patches a b c and m n q where at least one edge (including the vertices) of a b c intersects the plane where m n q is located, and the intersection point is inside m n q and also on the edge of a b c .
Definition 2.
Plane intersecting relationship: there are patches a b c and m n q , where the plane in which a b c lies intersects the plane in which m n q lies, and the intersection point is inside m n q and also on the extension line of the edge of a b c .
Definition 3.
Parallel relationship: there are patches a b c and m n q , and the plane in which a b c lies is parallel to the plane in which m n q lies.
Figure 4a–c shows 3D view images while Figure 4d–f shows front view images. We assume that the a b c is parallel to the horizontal plane, so it looks like a line in the front view images (Figure 4a–c). According to Figure 4a,d, the a b c and m n q have an intersecting relationship. The red points k and j are the intersection points of these two patches, and the red dotted line k-j is the intersecting line. The intersecting points and intersecting line are all inside two patches. According to Figure 4b,e, the a b c and m n q have a plane intersecting relationship. It means the plane in which a b c lies intersects the plane in which m n q lies. However, the intersecting points and intersecting line are only inside the a b c while outside the m n q , and intersecting points k and j are in the extension lines (blue dotted lines) of m-q and n-q respectively. According to Figure 4c,f, the a b c and m n q have a parallel relationship and the a b c is parallel to the m n q .
Based on these three definitions, two frames of the point cloud used for registration were meshed by using the greedy-projection-triangulation algorithm. Supposing that a b c was one patch of the first frame point cloud, m n q was the neighboring patch of a b c in the second frame and p m i d was the median plane of these two patches, the angle α t r i and distance d t r i were then calculated. If the sin 1 α t r i was less than 10 6 , these two patches were parallel, otherwise, the relation of these two patches needed to be computed using Möller’s theory [40]. In the intersecting or plane intersecting relationships, if α t r i was larger than α t r i m i n , each vertex of m n q was projected onto the p m i d forming a new patch m n q . However, in the plane intersecting relationship, the distance d p r o between the projection point and origin point was considered. If d p r o was bigger than the point moving threshold d p r o m a x , which meant that these two patches were not close enough, the projection operation was cancelled. Meanwhile, for the parallel condition, both a b c and m n q were projected onto the p m i d , forming two new patches, a b c and m n q . After projection, the distance d c e n between the geometric centers of two new patches was the basis for determining whether the projection operation was effective or not. If d c e n was larger than the threshold d c e n m a x , which meant that these two patches were not close enough, the projection operation was cancelled. However, the retrieval of the proposed neighboring patches was based on the k-d tree and patches’ geometric center, so d p r o was always less than d p r o m a x and d c e n was always less than d c e n m a x . Iteration produced the best result for two frames, and incremental registration optimization made all the frames into one. The detailed optimization algorithm is presented in Algorithm 2.
Algorithm 2: Registration optimization based on neighboring meshes
Input: different frames point clouds of plant after denoising and smoothing
Output: one frame of complete plant point cloud
Setting: global transformation matrix M g l o
(1) Registration: At the beginning, the first two frames are selected for processing. Fast global registration [41], which is more efficient than FPFH, is applied for rough registration, and ICP is applied for fine alignment, producing the temporary matrix M t e m p .
M g l o = M g l o M t e m p
(2) Meshing: Greedy projection triangulation is used to form triangular patches for these two frames.
(3) Searching neighboring patches: Calculating the patches’ geometric center and getting two center point clouds p c 1 and p c 2 . For each point in p c 1 , searching the neighboring points of selected point in p c 2 based on the k-d tree. The center point corresponds to the patch, so the neighboring patches of the patch a b c of p c 1 are a set T = { m n q i   |   i = 1 , 2 , 3 }
(4) Calculating the relationship between the patches: For each patch in set T, calculating the relationship between this patch and patch a b c . After projection, the new patch will take the place of the old one.
(5) Iteration and repetition: If α t r i is less than the minimum angle threshold α m i n , or d p r o is less than the minimum distance threshold d m i n , the optimization of these pairs of patches is completed. Repeating Steps 3–5 for all patches in the first frame.
(6) Down-sampling: After optimization, these two frames are combined into one frame, which is set as the new first frame. Due to repeated points, down-sampling is applied to reduce the point-cloud density.
(7) Applying to all frames: Taking the next frame from memory as the new second frame and then repeating the above operations until all frames are used.

3. Results and Discussion

The experiments were carried out on raw data obtained from 10 pots of oilseed rape. For each pot of plant, 10 frames of point cloud data from different views, which cover 360°and these data were processed by the proposed method to show the performance and robustness of proposed method.

3.1. Point-Cloud Noise Removal

At the beginning, there were approximately 210,000 points of raw date in each frame. Most of them were noise points, as shown in Figure 5a and Figure 6a. According to the definition in Section 2.1.1 and function (8), the red points in Figure 5b were valid points, and other points were noise points. The performance of removing BN points was evaluated by VPP. However, the perpendicular requirement, mentioned in Section 2.1.1, between the center axis of the plant and the camera-projection direction was not strictly satisfied (Figure 6a), the data still retained lots of BN points after using the pass-through filter directly (Figure 6c). As shown in Figure 6b, the point cloud data was rotated by MOBB to satisfy the perpendicular requirement, then BN points were removed more effectively by pass-through filter (Figure 6d). The comparison results between the above two methods for removing BN points in 10 frames of point cloud of one plant (plant 1) are presented in Table 1. In this experiment, the thresholds of the pass-through filter were (−9,40), (−25,50) and (35,70) cm in the X, Y and Z directions, respectively. These thresholds can preserve a more complete plant point cloud. Compared with the average VPP of the pass-through filter, which was 75.64%, the average VPP of the pass-through filter based on the MOBB was 92.05%. Several valid points removed by the method based on MOBB were mostly FPN points and accounted for a small number, so the method wouldn’t affect the quality of the point cloud. Table 2 shows the results of 10 pots of plants applied with above methods. The average VPP (AVPP), which was the average value of 10 frames’ VPP of each plant, was used. The AVPP still remained at a high level, with an average value of 92.28% and standard deviation (SD) of 2.27. According to the VPP and AVPP, the performance and the robustness of the proposed method was revealed by higher average value and smaller SD value, which indicated that the proposed method performed well with different frames of point cloud of a plant and different plants.
After removing BN points, there were still many ON and FPN points (Figure 7a,e). According to groups of experiments [33,42], the point cloud has good quality when r = 2 mm, K = 30, n = 2, and θ = 85°. As shown on the front view images in Figure 7a–d, all methods performed well in removing ON points. However, when it was switched to the side view, the results in Figure 7e–h indicated that there were significant differences between several methods. In Figure 7f, the data filtered by the radius-based outlier filter still contained many ON points around the leaves. As shown in Figure 7g,h, both the radius-density-based outlier filter and the proposed method generated relatively clean data. As mentioned in Section 2.1.1, FPN points existed at the edge of the leaf but were different from ON points, so the radius-density-based outlier filter could not deal with FPN points well. The proposed method got a better result by removing more FPN points on the boundary of the pot of the plant and ON points outside the leaves. As presented in Table 3, the radius-density-based outlier filter removed more ON points and had a higher average noise-reduction ratio (NRR) compared with the radius-based outlier filter. Further, considering the fact that the proposed method removed more FPN and ON points than the other two methods, it was reasonable that the proposed method reached average noise-reduction ratio of 14.06%. It was noteworthy that at the locations close to the boundary of leaves, the proposed method would mistake a few boundary points for FPN points and removed them from the point cloud, which brought a big SD of the noise-reduction ratio in Table 3. As for the whole plant, the proposed method showed comparable performance in removing ON and FPN points with high noise-reduction ratio (Table 4).
Above all, the proposed method performed well both in different frames of point cloud data of one plant and data of different plants. The small SDs from Table 1, Table 2, Table 3 and Table 4 indicated that the method had strong robustness.

3.2. Point-Cloud Smoothing

The smoothing effect of the bilateral filter mainly depends on σ c and σ s . The larger σ c , the smoother the point cloud was after processing. At the same time, the larger σ s , the more the point-cloud features that were preserved after processing. The optimal σ c and σ s were determined based on the different datasets acquired in this study. As shown in Figure 8, when σ c = 10 and σ s = 0.1, the distribution of the normals of the points was neat, which meant that the point cloud was smooth.

3.3. Optimization of Registration Based on Neighboring Meshes

The method proposed in this study was based on neighboring meshes, so the triangulation algorithm had a certain influence on the processing results. Meanwhile, the number of neighboring meshes processed also affected the results. If there are too many meshes, overlapping may occur. According to several sets of experiments, the optimization effect was best when α m i n = 20°, d m i n = 2 × the distance of the patches’ geometric center, the maximum number of iterations was 100, and the number of neighboring patches ≤ 3. Under this condition, 10 groups from different views covering 360° were tested, and each group had two adjacent frames of point cloud data. As shown in Table 5, the average Euclidean distance (AveEd) between parallel patches after optimization was 2.65 × 10−3 mm, and the average angle (AveAn) between intersecting and plane intersecting patches was 17.30°, which were 64.79% and 42.07% of these values before optimization, respectively. The smaller distance and angle indicated that the optimization method made neighboring patches from different frames of point cloud data become more appressed. The SDs of AveEd and AveAn after registration with optimization were low, which indicates that the optimization method had strong robustness. According to Table 6, the AveEd and AveAn were close to half of those value before optimization. The optimization method performed well in different plants stably with small SDs (Table 6).
From the above results, the proposed methods including the point-cloud noise removal method and the optimization method proved to have good performance and strong robustness not only in different frames of point cloud data of one plant but also different plants. Thus, we used 80 frames of data of one plant, which covered 360° to obtain a complete plant. 80 frames ensured a small angle between adjacent frames. Comparing Figure 9a with Figure 9b, the local layered points phenomenon improved. The leaf had three layers (in the red box of Figure 9a) before optimization, while it only had one layer (in the red box of Figure 9b) after optimization.

3.4. Efficiency

In order to obtain the point cloud data of a complete plant, we used 80 frames of point cloud data. In the tests of 10 pots of different plants, the average total time taken for the acquisition of the point cloud data of a complete plant was about 93.8 s, and the number of output plant points about one hundred thousand. Figure 10a illustrates the time consumed for each step in the proposed method. The longest time consumed by method was registration optimization based on neighboring meshes, which accounted for 64% of the total time consumed (Figure 10b). The calculation would be much faster if multi-thread processing was applied on a high configuration computer.

4. Conclusions

The plant 3D point-cloud optimization method proposed in this paper proved to be reliable for improving the quality of the plant point cloud. The point cloud was rotated into a better pose based on MOBB, and background noise points were totally removed with a pass-through filter, which preserved more valid points. For different plants, the method kept the valid point percent up to 92.28%, while that value was 82.24% only using pass-through filter. It was applicable to the plant-point-cloud data without plane objects due to the MOBB. The viewpoints and surface normals were effective in removing the outlier noise points and flying pixel noise points. In addition, we proposed applying neighboring mesh patches optimization during registration. After optimization, the average distance between the patches was 1.88 × 10−3 mm, and the average angle was 17.64°, which were 54.97% and 48.33% of those values before optimization, respectively. The impact of the layered-points phenomenon was effectively reduced, and the quality of the plant data were improved. The proposed method offers the potential to obtain complete and accurate plant data and may help to promote the popularization of plant-phenotyping research with low-cost sensors.

Author Contributions

Conceptualization, Z.M., D.S., Y.H. and H.C.; data curation, Z.M., H.X. and Y.Z.; formal analysis, Z.M.; methodology, Z.M.; writing—original draft, Z.M. and H.C.; writing—review & editing, Z.M., D.S., H.X., Y.Z., Y.H. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by National Natural Science Foundation of China (31801256), Synergistic Innovation Center of Jiangsu Modern Agricultural Equipment and Technology (4091600007), and Key R & D Program of Zhejiang Province, China (2020C02002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fiorani, F.; Schurr, U. Future Scenarios for Plant Phenotyping. Annu. Rev. Plant Biol. 2013, 64, 267–291. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Li, L.; Zhang, Q.; Huang, D. A review of imaging techniques for plant phenotyping. Sensors 2014, 14, 20078–20111. [Google Scholar] [CrossRef] [PubMed]
  3. Schurr, U.; Heckenberger, U.; Herdel, K.; Walter, A.; Feil, R. Leaf development in Ricinus communis during drought stress: Dynamics of growth processes, of cellular structure and of sink-source transition. J. Exp. Bot. 2000, 51, 1515–1529. [Google Scholar] [CrossRef] [Green Version]
  4. Dejong, T.M.; Da Silva, D.; Vos, J.; Escobar-Gutirrez, A.J. Using functionalstructural plant models to study, understand and integrate plant development and ecophysiology. Ann. Bot. 2011, 108, 987–989. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Paulus, S.; Dupuis, J.; Mahlein, A.K.; Kuhlmann, H. Surface feature based classification of plant organs from 3D laserscanned point clouds for plant phenotyping. BMC Bioinform. 2013, 14, 238. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Ivorra, E.; Sánchez, A.J.; Camarasa, J.G.; Diago, M.P.; Tardaguila, J. Assessment of grape cluster yield components based on 3D descriptors using stereo vision. Food Control 2015, 50, 273–282. [Google Scholar] [CrossRef] [Green Version]
  7. Nguyen, T.T.; Slaughter, D.C.; Maloof, J.N.; Sinha, N. Plant phenotyping using multi-view stereo vision with structured lights. In SPIE Commercial + Scientific Sensing and Imaging; SPIE: Baltimore, MA, USA, 2016; Volume 9866. [Google Scholar]
  8. Se, S.; Pears, N. 3D Imaging, Analysis and Applications; Pears, N., Liu, Y., Bunting, P., Eds.; Springer: London, UK, 2012; pp. 35–94. [Google Scholar]
  9. Rose, J.C.; Paulus, S.; Kuhlmann, H. Accuracy Analysis of a Multi-View Stereo Approach for Phenotyping of Tomato Plants at the Organ Level. Sensors 2015, 15, 9651–9665. [Google Scholar] [CrossRef] [Green Version]
  10. Xiang, L.; Tang, L.; Gai, J.; Wang, L. PhenoStereo: A high-throughput stereo vision system for field-based plant phenotyping-with an application in sorghum stem diameter estimation. In Proceedings of the 2020 ASABE Annual International Virtual Meeting, Lincoln, Nebraska, 13–15 July 2020. [Google Scholar] [CrossRef]
  11. Li, B.; An, Y.; Cappelleri, D.; Xu, J.; Zhang, S. High-accuracy, high-speed 3D structured light imaging techniques and potential applications to intelligent robotics. Int. J. Intell. Robot. Appl. 2017, 1, 86–103. [Google Scholar] [CrossRef]
  12. Rossi, R.; Leolini, C.; Costafreda-Aumedes, S.; Leolini, L.; Bindi, M.; Zaldei, A.; Moriondo, M. Performances Evaluation of a Low-Cost Platform for High-Resolution Plant Phenotyping. Sensors 2020, 20, 3150. [Google Scholar] [CrossRef]
  13. Paulus, S.; Schumann, H.; Kuhlmann, H.; Léon, J. High-precision laser scanning system for capturing 3D plant architecture and analysing growth ofcereal plants. Biosyst. Eng. 2014, 121, 1–11. [Google Scholar] [CrossRef]
  14. Sun, S.; Li, C.; Paterson, A.H. In-field high-throughput phenotyping of cotton plant height using LiDAR. Remote Sens. 2017, 9, 377. [Google Scholar] [CrossRef] [Green Version]
  15. Su, W.; Zhu, D.; Huang, J.; Guo, H. Estimation of the vertical leaf area profile of corn (Zea mays) plants using terrestrial laser scanning (TLS). Comput. Electron. Agric. 2018, 150, 5–13. [Google Scholar] [CrossRef]
  16. Del-Campo-Sanchez, A.; Moreno, M.; Ballesteros, R.; Hernandez-Lopez, D. Geometric characterization of vines from 3D point clouds obtained with laser scanner systems. Remote Sens. 2019, 11, 2365. [Google Scholar] [CrossRef] [Green Version]
  17. Liu, G.; Wang, J.; Dong, P.; Chen, Y.; Liu, Z. Estimating individual tree height and diameter at breast height (DBH) from terrestrial laser scanning (TLS) data at plot level. Forests 2018, 8, 398. [Google Scholar] [CrossRef] [Green Version]
  18. Malambo, L.; Popescu, S.C.; Murray, S.C.; Putman, E.; Pugh, N.A.; Horne, D.W.; Richardson, G.; Sheridan, R.; Rooney, W.L.; Avant, R.; et al. Multitemporal field-based plant height estimation using 3D point clouds generated from small unmanned aerial systems high-resolution imagery. Int. J. Appl. Earth Obs. Geoinf. 2018, 64, 31–42. [Google Scholar] [CrossRef]
  19. Paulus, S.; Behmann, J.; Mahlein, A.K.; Plümer, L.; Kuhlmann, H. Low-cost 3D systems: Suitable tools for plant phenotyping. Sensors 2014, 14, 3001–3018. [Google Scholar] [CrossRef] [Green Version]
  20. Soileau, L.; Bautista, D.; Johnson, C.; Gao, C.; Zhang, K.; Li, X.; Heymsfield, S.B.; Thomas, D.; Zheng, J. Automated anthropometric phenotyping with novel Kinect-based three-dimensional imaging method: Comparison with a reference laser imaging system. Eur. J. Clin. Nutr. 2016, 70, 475–481. [Google Scholar] [CrossRef]
  21. Cui, Y.; Schuon, S.; Chan, D.; Thrun, S.; Theobalt, C. 3D shape scanning with a time-of-flight camera. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1173–1180. [Google Scholar] [CrossRef] [Green Version]
  22. Corti, A.; Giancola, S.; Mainetti, G.; Sala, R. A metrological characterization of the Kinect V2 time-of-flight camera. Rob. Auton. Syst. 2016, 75, 584–594. [Google Scholar] [CrossRef]
  23. Knoll, F.J.; Holtorf, T.; Hussmann, S. Investigation of different sensor systems to classify plant and weed in organic farming applications. In Proceedings of the 2016 SAI Computing Conference (SAI), London, UK, 13–15 July 2016; pp. 343–348. [Google Scholar] [CrossRef]
  24. Vázquez-Arellano, M.; Reiser, D.; Paraforos, D.S.; Garrido-Izard, M.; Burce, M.E.C.; Griepentrog, H.W. 3-D reconstruction of maize plants using a time-of-flight camera. Comput. Electron. Agric. 2018, 145, 235–247. [Google Scholar] [CrossRef]
  25. Yang, H.; Wang, X.; Sun, G. Three-Dimensional Morphological Measurement Method for a Fruit Tree Canopy Based on Kinect Sensor Self-Calibration. Agronomy 2019, 9, 741. [Google Scholar] [CrossRef] [Green Version]
  26. Jiang, Y.; Li, C.; Paterson, A.H. High throughput phenotyping of cotton plant height using depth images under field conditions. Comput. Electron. Agric. 2016, 130, 57–68. [Google Scholar] [CrossRef]
  27. Xia, C.; Wang, L.; Chung, B.K.; Lee, J.M. In situ 3D segmentation of individual plant leaves using a RGB-D camera for agricultural automation. Sensors 2015, 15, 20463–20479. [Google Scholar] [CrossRef]
  28. Andújar, D.; Dorado, J.; Fernández-Quintanilla, C.; Ribeiro, A. An approach to the use of depth cameras for weed volume estimation. Sensors 2016, 16, 972. [Google Scholar] [CrossRef] [Green Version]
  29. Wang, X.; Singh, D.; Marla, S.; Morris, G.; Poland, J. Field-based high-throughput phenotyping of plant height in sorghum using different sensing technologies. Plant Methods 2018, 14, 1–16. [Google Scholar] [CrossRef] [PubMed]
  30. Hu, Y.; Wang, L.; Xiang, L.; Wu, Q.; Jiang, H. Automatic non-destructive growth measurement of leafy vegetables based on kinect. Sensors 2018, 18, 806. [Google Scholar] [CrossRef] [Green Version]
  31. Zhang, Z. Flexible camera calibration by viewing a plane from unknown orientations. Proc. IEEE Int. Conf. Comput. Vis. 1999, 1, 666–673. [Google Scholar] [CrossRef]
  32. Butkiewicz, T. Low-cost coastal mapping using Kinect v2 time-of-flight cameras. 2014 Ocean.-St. John’s Ocean 2014, 2015, 1–9. [Google Scholar] [CrossRef]
  33. Chunhua, X.; Ying, S. Obtaining and denoising method of three-dimensional point cloud data of plants based on TOF depth sensor. Trans. Chin. Soc. Agric. Eng. 2018, 34, 168–174. [Google Scholar] [CrossRef]
  34. Cheng, S.; Wu, W.; Yang, X.; Zhang, H.; Zhang, X. Rapid surfacing reconstruction based on Geomagic Studio software. Mod. Manuf. Eng. 2011, 1, 8–12. [Google Scholar] [CrossRef]
  35. Rosli, N.A.I.M.; Ramli, A. Mapping bootstrap error for bilateral smoothing on point set. AIP Conf. Proc. 2014, 1605, 149–154. [Google Scholar] [CrossRef] [Green Version]
  36. Han, X.F.; Jin, J.S.; Wang, M.J.; Jiang, W.; Gao, L.; Xiao, L. A review of algorithms for filtering the 3D point cloud. Signal Process. Image Commun. 2017, 57, 103–112. [Google Scholar] [CrossRef]
  37. Fleishman, S.; Drori, I.; Cohen-Or, D. Bilateral mesh denoising. ACM SIGGRAPH 2003 Pap. Int. Conf. Comput. Graph. Interact. Tech. 2003, 22, 950–953. [Google Scholar] [CrossRef]
  38. Rabbani, T.; Dijkman, S.; Van den Heuvel, F.; Vosselman, G. An integrated approach for modelling and global registration of point clouds. ISPRS J. Photogramm. Remote Sens. 2007, 61, 355–370. [Google Scholar] [CrossRef]
  39. Xiang, L.; Bao, Y.; Tang, L.; Ortiz, D.; Salas-Fernandez, M.G. Automated morphological traits extraction for sorghum plants via 3D point cloud data analysis. Comput. Electron. Agric. 2019, 162, 951–961. [Google Scholar] [CrossRef]
  40. Moller, T.; Trumbore, B. Fast, Minimum Storage Ray-Triangle Intersection. JGT 2012, 7651, 1–7. [Google Scholar] [CrossRef]
  41. Zhou, Q.-Y.; Park, J.; Koltun, V. Fast Global Registration Qian-Yi. Eur. Conf. Comput. Vis. ECCV 2016, 9906, 694–711. [Google Scholar] [CrossRef]
  42. He, D.; Shao, X.; Wang, D.; Hu, S. Denoising method of 3-D point cloud data of plants obtained by kinect. Trans. Chin. Soc. Agric. Mach. 2016, 47, 331–336. [Google Scholar] [CrossRef]
Figure 1. Acquisition system and point cloud acquisition. (a) Acquisition system; (b) the process of obtaining a single-frame point cloud.
Figure 1. Acquisition system and point cloud acquisition. (a) Acquisition system; (b) the process of obtaining a single-frame point cloud.
Sensors 21 00664 g001
Figure 2. The flow chart of the proposed method.
Figure 2. The flow chart of the proposed method.
Sensors 21 00664 g002
Figure 3. The relationship between the object, minimum oriented bounding box (MOBB) and coordinate in 2D space, (a) under ideal conditions and (b) under general conditions.
Figure 3. The relationship between the object, minimum oriented bounding box (MOBB) and coordinate in 2D space, (a) under ideal conditions and (b) under general conditions.
Sensors 21 00664 g003
Figure 4. Relationship between triangular patches, (ac) are 3D view images while (df) are front view images. (a,d) Intersecting patches. (b,e) Plane intersecting patches. (c,f) Parallel patches.
Figure 4. Relationship between triangular patches, (ac) are 3D view images while (df) are front view images. (a,d) Intersecting patches. (b,e) Plane intersecting patches. (c,f) Parallel patches.
Sensors 21 00664 g004
Figure 5. The valid points (red points). (a) Original point cloud. (b) Original point cloud with valid points.
Figure 5. The valid points (red points). (a) Original point cloud. (b) Original point cloud with valid points.
Sensors 21 00664 g005
Figure 6. Removal of BN points. (a) Original point cloud. (b) Point cloud rotated by MOBB. (c) Original point cloud after filtering with pass-through filter. (d) The rotated point cloud after filtering with pass-through filter.
Figure 6. Removal of BN points. (a) Original point cloud. (b) Point cloud rotated by MOBB. (c) Original point cloud after filtering with pass-through filter. (d) The rotated point cloud after filtering with pass-through filter.
Sensors 21 00664 g006
Figure 7. Results of different denoising methods: (ad) front view images, (eh) side view images. (a,e) The original point cloud. (b,f) The result after using the radius-based outlier filter. (c,g) The result after using the radius-density-based outlier filter. (d,h) The result after using the proposed method.
Figure 7. Results of different denoising methods: (ad) front view images, (eh) side view images. (a,e) The original point cloud. (b,f) The result after using the radius-based outlier filter. (c,g) The result after using the radius-density-based outlier filter. (d,h) The result after using the proposed method.
Sensors 21 00664 g007
Figure 8. The distributions of the normals of (a) the original points and (b) the points after smoothing.
Figure 8. The distributions of the normals of (a) the original points and (b) the points after smoothing.
Sensors 21 00664 g008
Figure 9. The results for the point cloud after registration: (a) Meshes of point cloud without optimization. (b) Meshes of point cloud after optimization.
Figure 9. The results for the point cloud after registration: (a) Meshes of point cloud without optimization. (b) Meshes of point cloud after optimization.
Sensors 21 00664 g009
Figure 10. (a) The time of each step in the proposed method. (b) The proportion of the time cost for each step.
Figure 10. (a) The time of each step in the proposed method. (b) The proportion of the time cost for each step.
Sensors 21 00664 g010
Table 1. The comparison results of two methods for removing background noise points from one plant (plant 1).
Table 1. The comparison results of two methods for removing background noise points from one plant (plant 1).
Frame NumberTotal Valid PointsPass-Through FilterPass-Through Filter Based on MOBB
Valid PointsTotal PointsVPP %Valid PointsTotal PointsVPP %
112,69012,68616,81975.4312,50913,47792.82
212,63812,63716,84175.0412,56613,58092.53
312,60012,59516,77575.081242713,61391.29
412,66912,65016,78075.3912,50213,66091.52
512,52612,50616,80174.4412,43613,62291.29
612,80712,78516,92575.5412,65913,76491.97
712,83712,80316,85475.9612,77313,78892.64
812,89312,85916,90276.0812,72613,82692.04
912,91112,87816,89576.2212,68913,81291.87
1013,09613,03316,88077.2112,86913,90492.56
Average\\\75.64\\92.05
SD \\\0.73\\0.54
Note: MOBB represents the minimum oriented bounding box. VPP represents the valid-point percent. SD represents the standard deviation.
Table 2. The average valid point percent of 10 pots of plants (%).
Table 2. The average valid point percent of 10 pots of plants (%).
MethodPlant Number
12345678910AverageSD
Method A75.6484.3187.9777.3084.0383.1982.9580.4082.9582.5782.243.37
Method B92.0589.6794.2897.1092.8094.2989.3790.5091.6891.0892.282.27
Note: Method A is pass-through filter method, and method B is pass-through filter based on MOBB method. MOBB represents the minimum oriented bounding box. VPP represents the valid-point percent. SD represents the standard deviation.
Table 3. The results of three methods for removing outlier noise and flying pixel noise points from one plant (plant 1).
Table 3. The results of three methods for removing outlier noise and flying pixel noise points from one plant (plant 1).
Frame NumberOriginal PointsRadius-Based Outlier FilterRadius-Density-Based Outlier FilterThe Proposed Method
PointsNRR/%PointsNRR/%PointsNRR/%
113,47713,3850.6812,7925.0811,10717.59
213,58013,4810.7312,7925.811,69113.91
313,61313,5160.7112,7796.1311,73713.78
413,66013,5340.9212,7726.511,72214.19
513,62213,4880.9812,7926.0911,69714.35
613,76413,6290.9812,9625.8311,78914.67
713,78813,6850.7512,9356.1911,76514.7
813,82613,6111.5612,9296.4911,79317.41
913,81213,6301.3212,9416.3111,40714.62
1013,90413,6981.4812,9756.6811,87114.13
Average\\1.01\6.11\14.94
SD\\0.33\0.46\1.39
Note: NRR represents Noise reduction ratio. NRR = (1 − point number after denoising/original point number) ∗ 100%. SD represents standard deviation.
Table 4. The noise reduction ratio (%) for 10 pots of plant from three methods.
Table 4. The noise reduction ratio (%) for 10 pots of plant from three methods.
MethodPlant Number
12345678910AverageSD
Method A1.01 1.27 1.82 2.49 2.53 0.58 1.05 1.97 3.46 1.82 1.80 0.87
Method B6.11 7.02 7.12 8.08 7.91 4.97 6.80 7.21 8.11 7.19 7.05 0.96
Method C14.94 9.95 10.15 10.86 12.54 9.78 10.78 11.16 10.98 11.25 11.24 1.52
Note: Method C is radius-based outlier filter method, method B is radius-density-based outlier filter MOBB method and method C is the proposed method. SD represents the standard deviation.
Table 5. The effects of registration optimization of one plant (plant 1).
Table 5. The effects of registration optimization of one plant (plant 1).
Test GroupBefore RegistrationAfter Registration without OptimizationAfter Registration with Optimization
AveEd
/ 10 3 mm
AveAn/°AveEd
/ 10 3 mm
AveAn/°AveEd
/ 10 3 mm
AveAn/°
13.9641.464.2041.093.5616.80
23.9441.464.3341.252.8517.19
33.7741.164.4140.722.8216.66
43.8641.494.1140.893.0716.85
54.2441.044.6340.732.9317.32
63.9241.574.0740.913.0117.37
73.4241.993.7141.042.8417.38
83.7441.863.9941.811.6017.74
93.1941.693.7241.341.3818.09
103.2341.433.7741.472.4517.58
Average 3.73 41.52 4.09 41.12 2.65 17.30
SD 0.34 0.290.31 0.35 0.67 0.44
Note: AveEd represents the average Euclidean distance of parallel triangles in the neighborhood, and AveAn represents the average angle of cross triangles or intersecting triangles in the neighborhood. SD represents standard deviation.
Table 6. The effects of registration optimization of 10 pots of plant.
Table 6. The effects of registration optimization of 10 pots of plant.
The Evaluation IndexPlant Number
12345678910AverageSD
A41.523737.0737.2337.3437.5638.3137.9537.7837.9837.971.32
B3.734.824.43.416.563.964.353.423.675.334.370.99
C41.1235.1535.9935.4635.836.5436.6736.0836.236.0136.501.68
D4.093.922.362.73.83.083.623.583.63.423.420.55
E17.317.717.2217.5718.1218.3217.4417.6417.4217.6417.640.35
F2.652.581.591.11.581.732.231.741.751.811.880.48
E/C (%)42.07 50.36 47.85 49.55 50.61 50.14 47.56 48.89 48.12 48.99 48.33 2.47
F/D (%)64.79 65.82 67.37 40.74 41.58 56.17 61.60 48.60 48.61 52.92 54.97 9.89
Note: A is the AveEd ( 10 3 mm) before registration; B is the AveAn (°) before registration; C is the AveEd ( 10 3 mm) after registration without optimization; D is the AveAn (°) after registration without optimization; E is the AveEd ( 10 3 mm) after registration with optimization; F is the AveAn (°) after registration with optimization. SD represents standard deviation.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, Z.; Sun, D.; Xu, H.; Zhu, Y.; He, Y.; Cen, H. Optimization of 3D Point Clouds of Oilseed Rape Plants Based on Time-of-Flight Cameras. Sensors 2021, 21, 664. https://doi.org/10.3390/s21020664

AMA Style

Ma Z, Sun D, Xu H, Zhu Y, He Y, Cen H. Optimization of 3D Point Clouds of Oilseed Rape Plants Based on Time-of-Flight Cameras. Sensors. 2021; 21(2):664. https://doi.org/10.3390/s21020664

Chicago/Turabian Style

Ma, Zhihong, Dawei Sun, Haixia Xu, Yueming Zhu, Yong He, and Haiyan Cen. 2021. "Optimization of 3D Point Clouds of Oilseed Rape Plants Based on Time-of-Flight Cameras" Sensors 21, no. 2: 664. https://doi.org/10.3390/s21020664

APA Style

Ma, Z., Sun, D., Xu, H., Zhu, Y., He, Y., & Cen, H. (2021). Optimization of 3D Point Clouds of Oilseed Rape Plants Based on Time-of-Flight Cameras. Sensors, 21(2), 664. https://doi.org/10.3390/s21020664

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop