1. Introduction
Grain is a common type of cargo for transportation. Due to its fluidity, bulk transportation of grain can better utilize the space of transport vehicles and improve loading and unloading efficiency. Therefore, in practice, grain is usually transported in bulk. The loading site for bulk grain cargo is shown in
Figure 1. The grain storage bins open the grain pipeline valve, using gravity to directly pour the grain from the discharge port into the transport vehicle. When the outflowing grain reaches the set weight, the valve is closed to stop the discharge port. In this process, to fully utilize the vehicle’s space and prevent the grain from overflowing excessively, the loading workers must climb to a high position to observe the loading situation inside the vehicle and notify the driver to adjust the vehicle’s position. However, due to the characteristics of bulk grain, a large amount of dust is inevitably generated at the loading site, and the loading equipment produces noise during operation. This makes it difficult for loading workers to observe the grain’s condition inside the vehicle and communicate with the driver using traditional methods. This process results in high labor intensity for the loading workers, significant safety hazards over long-term work, and low grain-loading efficiency. Therefore, it is worth researching and exploring how to automatically guide the driver to adjust the vehicle’s position, reduce the labor intensity of the loading workers, eliminate safety hazards during the loading process, and ultimately improve overall loading efficiency.
In recent years, with the development of deep learning and perception technologies, sensors have been widely applied in various important fields such as transportation, logistics, robotics, and industrial production [
1,
2,
3]. Using sensors to obtain scene information, and performing target recognition or semantic segmentation based on perceptual data, has become the key to achieving automation and intelligence [
4,
5]. To address the aforementioned issues, suitable sensors can be selected to acquire bulk grain-loading information, extract the vehicle’s position and the grain’s shape inside the vehicle, recognize the bulk grain-loading status, guide the vehicle to adjust its position based on the loading status, and achieve automated grain loading.
In practical engineering, bulk grain-loading tasks need to be carried out according to production schedules, with timing that is not fixed but is influenced by weather conditions. At the loading site when there are no loading tasks, staff or working vehicles will also pass through the loading area. Their main modes of passage include walking, cycling, and driving. Obviously, these unrelated individuals must not affect or interfere with the recognition of the bulk grain-loading status. To recognize the bulk grain-loading status, it is necessary to identify the vehicle that is to be loaded, determine the vehicle’s position, and perceive the shape of the grain and the vehicle’s information. Cameras find it difficult to obtain precise depth information and are greatly affected by lighting conditions. If cameras are used for perception, it is challenging to accurately capture vehicle and grain data during the loading process, and additional lighting equipment is needed for nighttime operations. LiDAR, on the other hand, can directly generate three-dimensional point cloud data with high measurement accuracy and robustness, and it is suitable for long-distance and nighttime work. This makes LiDAR applicable in many environments where cameras struggle to function effectively [
6]. The three-dimensional point cloud data from bulk grain loading, which includes precise vehicle position information and grain shape data, forms the basis for recognizing the loading status.
Deep learning technology, as an efficient representation learning algorithm, can extract key features from these point cloud data to perform complex object-detection and -segmentation tasks [
7,
8]. PointNet [
9] has set a milestone in the deep learning processing of point clouds, directly extracting global features from point sets, but it has encountered difficulties in capturing local geometric details. PointNet++ [
10] introduces a hierarchical network structure that can capture hierarchical features from local to global, significantly improving the processing capability for complex point cloud data and fine local structures. Building on this, PointNeXt [
11] has further enhanced performance and efficiency by improving training strategies, introducing separable MLP (Multilayer Perceptron), and incorporating inverted residual MLP. These advancements allow the network to learn more complex features and representations from point cloud data, leading to better generalization and robustness in various point cloud-processing tasks. DGCNN [
12] uses a graph convolutional network to process point clouds, effectively capturing local features through a dynamic graph structure, and LDGCNN [
13] optimizes the network structure by connecting hierarchical features of different dynamic graphs, effectively addressing the gradient vanishing problem. Shellnet [
14], with its statistical data from concentric spherical shells, effectively resolves the issue of order ambiguity in point cloud data, maintaining high performance with a relatively small number of model layers. PointMLP [
15] introduces a pure residual network equipped with lightweight geometric affine modules, significantly enhancing the speed and accuracy of point cloud processing.
Although the above methods can all accomplish point cloud-classification and -segmentation tasks, their different processing methods mean that their application scenarios also vary. Convolutional neural networks processing point clouds are often used for identifying and separating complex structures [
16]. They are widely applied in semantic segmentation tasks in urban road areas, 3D mapping fields, and medical fields [
17,
18,
19]. Directly processing point clouds is more computationally efficient and more suitable for object detection and segmentation in the automation field [
20], such as recognizing human actions [
21], segmenting object parts, and measuring target volumes [
22]. The bulk grain-loading process is a fixed scene for identifying and segmenting specific targets, where the collected point cloud data types are relatively few, and there is no need for overly complex model structures to recognize multiple types of targets. As a classic network that directly processes point clouds, PointNet++ is more suitable for the grain-loading task due to its lower parameter count and faster training speed. However, the point clouds in the bulk grain-loading process have characteristics such as feature loss due to grain or dust occlusion, a large number of points, complex processing, and sparse points at intersections and edges. Therefore, we propose a novel method tailored for the bulk grain-loading process, which is based on the PointNet++ network and adjusted to address the unique challenges of this application by incorporating specific characteristics of bulk grain loading.
Based on the above context, this paper conducts research on the method for recognizing the bulk grain-loading status based on LiDAR. Using dual LiDAR sensors for joint perception ensures complete coverage of the loading area. Data preprocessing is carried out to eliminate interference points and fill in occluded parts. A bulk grain-loading status-recognition network, PNGL (Point Net Grain Loading), is constructed to classify the target point cloud, extract high-dimensional features of the vehicle, and achieve vehicle target detection, thereby avoiding missed and false detections. The vehicle components are finely segmented to obtain vehicle information and grain shape, and the grain-loading status is determined. The current status and prompt information are output to assist operators in loading and unloading tasks, promoting the automation of bulk grain loading. This research effectively addresses the shortcomings of traditional judgment methods, reduces the labor intensity of operators, eliminates safety hazards, improves the efficiency of bulk grain loading, and has significant importance for achieving automated loading.
3. Method for Recognizing Bulk Grain-Loading Status
3.1. Sensor Installation
To solve the problem of limited perception during bulk loading, it is necessary to reasonably plan the installation positions of the sensors. The principle is to ensure that the sensors are easy to install and reproducible without affecting their performance. The sensor installation positions in this paper are shown in
Figure 3. Specifically, LiDAR sensors are installed on both sides of the discharge port, 30 cm away from it, and a perception coordinate system is established. The origin is the ground directly below the discharge port, with the vehicle’s forward direction as the
x-axis, the vertical upward direction as the
y-axis, and the left-to-right direction as the
z-axis.
As shown in
Figure 3, increasing the installation height of the LiDAR sensors to expand the perception range would cause significant occlusion of the LiDAR’s field of view by the discharge port. Using higher line beam and larger range LiDAR sensors would lead to incomplete perception due to occlusion by the vehicle’s walls. Based on the above analysis, we chose to install the LiDAR sensors at the fixed positions shown in
Figure 3 to collect point cloud data of the vehicles and grain in the loading area.
3.2. Sensor Calibration and Fusion
During the installation of LiDAR sensors, variations in fixed components and installation methods can lead to slight tilts in the LiDAR, as shown in
Figure 4. The tilt angle can cause discrepancies in the actual position of targets, affecting subsequent recognition and judgment. To ensure high-quality point cloud data collection and achieve LiDAR data synchronization, it is necessary to calibrate the LiDAR sensors to eliminate tilt angle errors.
For LiDAR calibration, this study references Wen’s method [
25], utilizing an adaptive approach for point cloud calibration. The advantage of this method is that it maintains an error within 0.1 degrees, does not require a calibration board, and is more suitable for the installation and operation of multiple devices in industrial scenarios. The specific method is as follows: Perceive the loading area without any vehicles to obtain a dense ground point cloud. Use the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) algorithm to fit and separate the complete ground point cloud. Calculate the angle between the ground point cloud and the ideal coordinate system. Transform the original point cloud using the corresponding rotation and translation matrices in three directions, as shown in Equation (2), to obtain the point cloud data in the target coordinate system.
where
α is the tilt angle in the
x-axis direction,
β is the tilt angle in the
y-axis direction, and
γ is the tilt angle in the
z-axis direction.
As shown in
Figure 5, after calibration, the point clouds obtained by the two LiDARs do not match the target point cloud in the loading site coordinate system. To ensure that the point cloud matches the real target, it is necessary to perform translation and rotation transformations on the target point cloud to achieve data fusion.
For the translation transformation, since the LiDAR positions are fixed, the translation vector in the three directions can be calculated by directly measuring the difference between the target point cloud and the actual position of the target.
For the rotation transformation, since the LiDARs have already been calibrated, it is only necessary to convert the LiDAR coordinate system to the loading site coordinate system. Specifically: For the left LiDAR, reverse its z-axis, with the rotation vector being. For the right LiDAR, reverse its x-axis, with the rotation vector being. By applying these transformations, the coordinate systems can be unified and data fusion can be achieved.
3.3. Data Preprocessing
Point cloud data are subject to noise, duplicate points, and outliers due to factors such as the sensor’s installation angle, detection range, and dust interference. If left unprocessed, these issues can adversely affect recognition performance. To eliminate irrelevant points and enhance system performance, this study employs pass-through filtering and statistical filtering to remove interference points.
Pass-through filtering is often used for initial data processing, allowing for the direct extraction of point cloud data within a specified range. After pass-through filtering, the point cloud coordinates within a length of 25 m, a width of 6 m, and a height of 6 m are obtained.
Statistical filtering can effectively remove outliers in the point cloud, eliminating the interference caused by dust during the bulk grain-loading process and ensuring the effectiveness of model training. The specific process is shown in Equation (3), where
μ represents the average distance from a certain point to all other points,
n is the total number of points,
σ denotes the standard deviation of the distances, and
S indicates a custom standard deviation coefficient. Using a distance threshold
Dmax as a reference the average distance between each point and its 8 nearest neighbors is calculated. If this distance exceeds the threshold, the point is considered an outlier. In this study, through multiple experiments, the parameters have been set such that the number of nearest neighbors is 8, and
S is set to 1 for statistical filtering. The results are shown in
Figure 6.
3.4. Dataset Construction
This study conducted field data collection at a grain terminal in a port located in southeastern China. The primary types of grain at this terminal include wheat, corn, and soybean meal. The terminal has nearly one hundred grain-loading and -unloading platforms, with an annual throughput exceeding 20 million tons, highlighting a significant and urgent need for automated grain handling.
The main vehicle types used for grain transport at this port are large trailers and medium-sized trucks. When there are no loading tasks, the bulk grain-loading area is also used for vehicle and pedestrian passage, where pedestrians can move either on foot or by bicycle. Two LiDAR sensors were installed at the positions shown in
Figure 3, and point cloud data of vehicles and pedestrians passing through the monitored area were collected multiple times. In total, 3195 samples of vehicles and pedestrians under various conditions were gathered.
Subsequently, the collected point cloud data were manually classified and labeled to form a bulk grain-loading dataset. To minimize discrepancies due to different classification standards and segmentation boundaries among different individuals, the labeling was conducted by the same person using the open-source software CloudCompare V2.12.4.
After processing the collected data through fusion, classification, and cleaning, a total of 2810 samples were obtained for the point cloud-classification dataset and the component-segmentation dataset. The point cloud-classification dataset was divided into five categories: pedestrians, cars, bicycles, trucks, and trailers. The results are shown in
Figure 7, where the point clouds have been normalized and centered to ensure visualization clarity. The segmentation dataset includes two categories: trucks and trailers, with each category further divided into eight components: truck cab, front wall of the truck body, interior of the truck body, rear wall of the truck body, trailer cab, front wall of the trailer body, interior of the trailer body, and rear wall of the trailer body. The manually labeled components of the vehicles are shown in
Figure 8, and the data for each part are listed in
Table 2.
3.5. PNGL Network Design
Based on the collected data, it is observed that the bulk grain-loading point cloud has certain characteristics such as partial occlusion, a large number of points, dust interference when grain is falling, and sparse edges of the vehicle body. To accommodate these characteristics, we propose an improved PointNet++ network named PNGL (Point Net Grain Loading).
The PNGL network adopts octree sampling, which provides faster processing speeds and better preservation of spatial features of the point cloud, thereby reducing dust interference. Additionally, since the point clouds of different vehicle types and different batches of loading are quite similar, and the vehicle point clouds are mobile during the loading process, the network employs global average pooling to reduce the risk of overfitting, decrease sensitivity to positional information, and extract global features.
The structure of the PNGL network is shown in
Figure 9. It consists of three main parts: feature extraction, point cloud classification, and component segmentation.
Feature Extraction: This part includes three set abstraction layers to recursively extract multi-scale features at the scale of {1/4, 1/64, 1/256} for the input point cloud consisting of
N points. The raw point cloud data is processed through these layers, transforming it into high-dimensional nonlinear representations and aggregating features within both local and global regions. Specifically, the first set abstraction layers take an
N × 3 matrix as the input and outputs an N/4 × 64 matrix of N/4 subsampled points with 64 dimensional feature vectors summarizing the local contextual information. Following the same principle, the second and third set abstraction layers proceed in the same manner, ultimately yielding an N/256 × 512 matrix.
Figure 10 shows an example set abstraction layers.
In summary, the PNGL network is specifically designed to handle the unique challenges presented by bulk grain-loading point clouds, including occlusions, large datasets, dust interference, and sparse edges. By employing octree sampling and global average pooling, along with set abstraction and feature propagation layers, the PNGL network effectively extracts and processes high-dimensional features to achieve accurate point cloud classification and component segmentation.
The set abstraction layer consists of three sub-layers: the Sampling Layer, the Grouping Layer, and the PointNet Layer. As shown in
Figure 10, the process is as follows:
Sampling Layer: The point cloud data is input, and octree sampling selects a set of sampled points N′. These points define the centroids of local regions.
Grouping Layer: Using a query ball, this layer constructs local region sets by finding K points within the radius of a sphere around each centroid. The output point set at this stage is N′ × k × (d + C), where K represents the number of points in the neighborhood of each centroid point. Each set abstraction module has different values for the sampling number K and the sampling radius R. Additionally, K and R increase with each layer, allowing the set abstraction to capture local point cloud features at different scales. Multiple set abstractions output the global features of the point cloud.
PointNet Layer: This layer uses a small PointNet network to encode the local region patterns into feature vectors. It employs a multi-layer perceptron (MLP) structure and applies global average pooling to the features. The resulting point set after feature extraction by the PointNet layer is N′ × (d + C′).
The detailed operations within each sub-layer are as follows: Sampling Layer: Selects a subset of points from the input point cloud using octree sampling, which efficiently captures the structure and density variations in the point cloud. Grouping Layer: For each sampled point (centroid), this layer groups neighboring points within a certain radius to form local regions. The radius and the number of neighbors increase with each subsequent layer, allowing the network to capture features at varying scales. PointNet Layer: Each local region is passed through a small PointNet network. This involves applying shared MLPs to each point in the local region, followed by max pooling to obtain a feature vector that represents the entire region. Global average pooling is then applied to aggregate these features, resulting in a compact representation of the point cloud’s global features.
By combining these three sub-layers, the set abstraction layer effectively captures and encodes local and global features of the point cloud. This multi-scale feature-extraction process is crucial for accurately classifying and segmenting point clouds in the context of bulk grain loading, where point clouds are characterized by occlusions, dust interference, and varying densities.
For the component-segmentation task, the network needs to progressively propagate the extracted high-dimensional features back to the original point set to ultimately obtain the class label for each point. This method decodes the features through a three-layer feature propagation mechanism, as shown in
Figure 11.
The process of feature propagation is as follows: To propagate the point features from
N × (
d +
C) to
N′ this study uses inverse distance weighted interpolation [
10] (IDW) as shown in Equation (6), where
p is set to 2 and
m is set to 3. The feature values at the coordinates of
N′ are interpolated from the features of
N to achieve feature propagation. These interpolated features are then concatenated with the previous layer’s features and passed through a MLP with shared fully connected layers and activation functions to update the feature vector of each point. This process is repeated until the features are propagated to the original point set, ultimately achieving point cloud segmentation.
where
f is the eigenvalue,
p denotes the degree of influence of distance on weights, and
m denotes the interpolation calculation by taking
m points in the known point set.
3.6. Recognition of Loading Status
The aforementioned model is applied to the bulk grain-loading process for the classification and segmentation of target vehicles. By calculating the maximum, minimum, and average coordinates of the point clouds of each segmented part, vehicle data can be obtained. The loading status of the bulk grain can then be determined based on the acquired vehicle data, as shown in
Figure 12. Currently, the loading status are divided into the following five categories:
Empty Vehicle: The average height of the compartment is close to the height from the bottom of the compartment to the ground, approximately the minimum value.
Grain Loading: The height of the point cloud in the compartment below the discharge port gradually increases.
Overheight Warning: The height of the point cloud in the compartment below the discharge port is about to exceed the height of the front wall of the compartment.
Loading Complete: The average height inside the compartment is close to the height of the compartment walls, and the rear wall of the compartment is near the discharge port.
Standby: There is no target vehicle in the loading area, and the system is in standby mode.
Based on the status outputs, the following instructions are executed:
Empty Vehicle: Calculate the vehicle’s relative position and direct the vehicle to move so that its front wall slightly exceeds the discharge port.
Grain Loading: Calculate the grain height in the discharge port area and the grain volume, compute the loading percentage, and provide real-time feedback to the driver. When the grain height is about to exceed the height of the front wall of the compartment, transition to the Overheight Warning status.
Overheight Warning: Prompt the driver to move the vehicle forward until the grain height in the discharge port area returns to a safe range.
Loading Complete: Once loading is complete, stop the discharge from the port and prompt the vehicle to leave the loading area. Standby Status: The system waits for the next vehicle to enter the loading area.
By implementing these steps, the system can accurately identify the current status of grain loading and provide appropriate instructions to ensure safe and efficient loading operations. The real-time feedback and automated instructions reduce manual intervention, improve loading efficiency, and enhance safety during the bulk grain-loading process.
4. Experiments
To evaluate the performance of the PNGL network in recognizing bulk grain-loading status, multiple experiments were conducted on the collected bulk grain-loading dataset.
4.1. Experimental Setup
To ensure training quality and increase model robustness, data augmentation techniques such as random jittering and random flipping were applied to part of the collected data. The hyperparameters for the model were set as follows: batch size of 16,200 epochs, initial learning rate of 0.001, learning rate decay of 0.5, with decay applied every 20 iterations, and using the AdamW optimizer. After 200 epochs of iterative training, the best segmentation model was used to segment the data in the test set.
4.2. Evaluation Metrics
The following evaluation metrics were used to quantitatively compare and analyze the classification and segmentation results of the bulk grain-loading point clouds: Overall Accuracy (OA), Mean Accuracy (mAcc) [
26], and Mean Intersection over Union (mIoU) [
27].
where
C is the total number of classes,
TPi is the number of true positives for class
i,
TNi is the number of true negatives for class
i,
FPi is the number of false positives for class
i,
FNi is the number of false negatives for class
i.
4.3. Results of Experiments
To evaluate the point cloud-classification performance of our network, we calculated the prediction probabilities for point cloud classification and segmentation and plotted the corresponding heatmap, as shown in
Figure 13. From the figure, it can be seen that for the point cloud-classification task, our method achieved an Overall Accuracy (OA) of 97.9%. Specifically, the classification results for pedestrian and bicycle point clouds were the best, mainly because pedestrian and bicycle targets are smaller and have significantly different features compared to vehicle point clouds, resulting in optimal classification performance. The classification accuracies for the two main targets, trucks and trailers, were 94.4% and 97.6%, respectively, generally meeting the task requirements. However, we observed some classification errors, which may be due to grain loaded inside the cargo area obstructing part of the vehicle features, leading to misclassification. To address this issue, we plan to add a verification procedure in the future to quickly check the classified point clouds and further improve accuracy.
The point cloud segmentation predictions for the two research objects are shown in
Figure 14a,b. For the segmentation of truck components, the overall performance accuracy is relatively good, with segmentation errors concentrated between the body and the front and rear walls. This is because the increased height of the grain pile covers the cargo walls, making them difficult to distinguish. However, the extraction of vehicle data is completed in the empty load status, so it only causes the grain point cloud extraction to be wider than it actually is. Therefore, the aforementioned phenomenon is within a reasonable range and does not affect the effectiveness of our method. Similarly, the trailer also has this issue, but because the rear half of the trailer is not loaded too high, the distinction between the trailer body and the rear wall is better.
4.4. Comparative Experiments
To fairly demonstrate the robustness and effectiveness of our architecture, we compared it with other representative point cloud models, including PointNet [
9], PointNet++ [
10], ShellNet [
14], and PointMLP [
15]. Additionally, we incorporated metrics for FLOPs (Floating Point Operations) and Params (Parameters) to assess the complexity and scale of the model’s parameters, ensuring that the model is practical and usable in engineering applications. The test results for the two tasks are shown in
Table 3 and
Table 4.
In the point cloud-classification task, the PointNet network exhibited the lowest classification accuracy. This phenomenon is attributed to the fact that PointNet only extracts global features, neglecting the spatial adjacency relationships between points, which contain some key feature information. PointNet++, with its hierarchical structure, can extract local point information, thus achieving better point cloud-classification results. However, due to the similar geometric structures at the edges, feature aggregation through local neighborhoods can lead to overly smooth feature transitions. Moreover, when grain obstructs the vehicle walls, it causes edge sparsity and results in the loss of local details. Shellnet assigns ordered convolutional weights to the unordered point cloud by statistical features within concentric spherical shells, but loses local details and high-frequency information when aggregating features from each shell layer, leading to inferior performance in bulk grain-loading tasks compared to PointNet++. In contrast, PointMLP recursively aggregates local features through a feed-forward residual MLP network, demonstrating superior performance, but with higher FLOPs and Params, consuming more computational resources. Overall, our network, improved specifically for the characteristics of bulk grain loading, achieved an OA of 97.9% and a mAcc of 98.1% in the point cloud-classification task, offering high accuracy with a simple model, making it more suitable for practical engineering applications.
In the point cloud-segmentation task, these models showed similar characteristics.
Figure 15 shows the comparison of point cloud-segmentation results between this method and the PointNet++ network. For the PointNet++ network, it can be observed that there are classification errors at the edges of components and the junctions between the vehicle walls and the cargo area. These errors are mainly due to dust interference and the occlusion of vehicle components by grain, leading to similar structures or feature loss. For the proposed network in this paper, the results closely match the actual situation. It correctly classifies the cab, front wall, rear wall, and cargo area for both types of vehicles. Even when grain accumulates and obscures the cargo area walls, it correctly segments the corresponding components. This is primarily because the proposed method includes point cloud filtering, which reduces dust interference. It also uses octree sampling and a weighted cross-entropy loss function, which better preserves local details, reduces the weight of lost features, and addresses the classification imbalance caused by the characteristics of grain and environmental factors. This resulted in an OA of 99.1% and a mIOU of 96.6%. In summary, our method maintains high accuracy in target recognition and vehicle component segmentation during the bulk grain-loading process, making it more suitable for determining the bulk grain-loading status.
The above results indicate that this method maintains high accuracy for target recognition and vehicle component segmentation during the bulk grain-loading process, making it more suitable for determining the bulk grain-loading status.
5. Discussion
Currently, research on recognizing the bulk grain-loading status is still in its infancy. This paper utilizes LiDAR sensors for data collection and relies on deep learning networks to accurately detect and classify vehicles and their loading status, achieving the recognition of bulk grain-loading statuses and filling a gap in this field.
Experimental results demonstrate that the proposed PNGL network has achieved high accuracy in both point cloud-classification and component-segmentation tasks. To provide a comprehensive assessment, our results were compared with several state-of-the-art deep learning methods for point clouds, including PointNet [
9], PointNet++ [
10], Shellnet [
14], and PointMLP [
15]. In the point cloud-classification task, our PNGL network achieved an OA of 97.9% and a mAcc of 98.1%. By effectively capturing both global and local features, it surpassed these advanced methods even in the challenging environment of bulk grain-loading tasks characterized by occlusions and dynamic point clouds. In the component-segmentation task, it achieved a segmentation accuracy of 99.1% OA and 96.6% mIOU. It demonstrated excellent segmentation performance, especially in differentiating components in close proximity and handling partial occlusions caused by grains and dust, while maintaining a lightweight structure that reduces computational resource consumption, making it more suitable for bulk grain-loading tasks.
This study is dedicated to solving the problems of high labor intensity, poor safety, and low loading efficiency in the bulk grain-loading process. Due to the particularity of bulk grain-loading tasks, occlusions caused by loading and unloading equipment and vehicle body components leading to incomplete point clouds, as well as interference from dust, pose challenges to the task of recognizing grain-loading status. Our study addresses several key issues related to the bulk grain-loading process:
How does the PNGL network handle occlusions and dust interference compared to other methods? The PNGL network, combined with octree sampling and advanced feature propagation techniques, enhances its ability to filter out noise and maintain high precision even in dusty and occluded environments. The improvement in performance metrics compared to other methods proves this point.
What are the advantages of using the PNGL network in dynamic and sparse point cloud scenarios? The dynamic nature of grain loading presents challenges for point cloud processing. The hierarchical feature extraction and global average pooling of the PNGL network reduce sensitivity to sparsity and motion in point clouds, thereby achieving more reliable classification and segmentation results.
Why is the PNGL network more suitable for bulk grain-loading applications? The bulk grain-loading process requires precise detection of load levels to prevent overflow and optimize space utilization. The PNGL network can accurately segment vehicle components and detect grain shapes, ensuring that it can provide real-time feedback for efficient and safe loading operations.
We address research questions by associating the deep learning network with the initial goals of improving loading efficiency, reducing labor intensity, and assisting in the automation of loading vehicles. Specifically, the proposed method effectively detects and segments common transport vehicle types, including the cab, front and rear walls of the cargo area, the interior of the cargo area, and the grain. By analyzing the interior of the cargo area, the method recognizes the grain shape, and by assessing the relative positions of vehicle components, it detects the vehicle’s position. Compared to traditional manual detection and other deep learning models, this method has the following advantages:
Reliable Loading Status Recognition: By comparing the detected grain shape and height with predefined thresholds and setting buffer zones, the method ensures that all four vehicle statuses can be correctly determined. This prevents accidents caused by grain height exceeding the cargo area’s limits due to occasional recognition errors.
Intelligent Prompting: Based on the detected status and grain shape, the system outputs vehicle movement instructions to the driver, effectively guiding the loading process. This significantly reduces the labor intensity of workers, improves safety, and enhances loading efficiency.
Higher Recognition Accuracy: Compared to commonly used deep learning models, this method shows significant improvements in accuracy for both point cloud-classification and -segmentation tasks.
Robustness in Complex Environments: This method does not require large-scale modifications to the loading site. It can accurately detect vehicle positions and recognize loading statuses even in the presence of dust interference and occlusion in the point cloud data.
6. Conclusions
This paper addresses the issues present in traditional bulk grain loading and proposes a LiDAR-based method for recognizing bulk grain-loading status. This method involves placing two LiDAR sensors in the loading area to acquire point cloud data. Through data fusion and preprocessing, a complete vehicle point cloud is obtained. Based on the characteristics of the bulk grain-loading process, a deep learning network architecture, PNGL, is designed for vehicle recognition and component segmentation. This enables the detection of vehicle positions, perception of grain shapes, and recognition of bulk grain-loading status. Finally, the method outputs prompt instructions based on the recognized grain-loading status, effectively solving the communication difficulties, safety risks, and inefficiencies present in traditional bulk grain-loading processes.
In future research, we plan to further streamline the model structure to ensure performance while reducing hardware dependency and costs. We aim to collect data from a wider variety of grain-loading scenarios to enrich the dataset, enhance the model’s generalization ability, robustness, and stability, and further promote the development and application of automated bulk grain-loading technology.