1. Introduction
In recent years, Japanese agriculture has faced increasing challenges, particularly due to the declining number of core agricultural workers and the growing proportion of elderly workers. These demographic trends, influenced by the nation’s declining birthrate and aging population, have significantly affected the agricultural workforce [
1]. Concurrently, heightened environmental awareness has brought organic farming into the spotlight. Organic farming not only mitigates the risk of environmental pollution, but also enhances the value of agricultural products [
2,
3,
4,
5]. However, a major challenge in organic farming is the increased labor intensity required for weeding, which is more demanding compared to conventional herbicide usage [
6]. Robotic automation has been identified as an effective solution to this issue [
7]. Among the various methods for automating weed control in organic farming, weeding cultivators are particularly notable [
8]. These devices are used in the cultivation of several crops, including legumes, corn, potatoes, onions, and sugar beets. The operation of weeding cultivators requires delicate precision, often at the centimeter level, over extended periods, which is physically and mentally demanding, further emphasizing the need for automation.
Automating weeding cultivator operations requires systems capable of both following crop rows and recognizing these rows accurately. In actual field conditions, the transplanting of crops—whether mechanically or manually—can result in varying degrees of row misalignment. Consequently, weeding machines without feedback mechanisms cannot automatically adjust to the curves of the crop rows, which may result in crop damage. Therefore, it is essential for automated systems to avoid damaging seedlings. Weeding cultivators are commonly employed in organic onion cultivation [
9]. The row spacing in onion farming typically ranges from 12 to 15 cm, which is narrower than that of crops such as corn (20–70 cm) and soybeans (60–70 cm). As such, a high-precision tracking performance relative to the crop rows is required [
10,
11,
12]. This precision is crucial for successful automation. Research on automated steering systems for agricultural machinery has explored methods utilizing both Global Navigation Satellite Systems (GNSSs) and camera-based systems [
13,
14]. In the case of automatic steering for tractors, hydraulic steering systems have been developed that can track target lines with a steering angle deviation of less than 2.192° and a maximum lateral path tracking error of 4.39 cm [
15,
16]. Using GNSSs and Inertial Navigation Systems (INSs), agricultural tractors have been able to achieve a lateral path tracking error as low as 2.94 cm under straight path conditions [
17]. However, GNSS-based navigation systems are vulnerable to environmental factors such as weather conditions, satellite positioning, and surrounding obstacles. To overcome these limitations, vision-based road detection systems independent of GNSSs are being developed. These systems work by segmenting road surfaces and edges from RGB images captured by cameras. The resulting lateral error from the machine vision system can then be transmitted to the tractor’s automatic navigation system, enabling automatic driving with lateral errors of less than 0.2 m on unpaved roads and less than 0.4 m on paved roads [
18]. This highlights that automatic weeding cultivator operations are feasible if crop rows can be effectively recognized.
Crop row recognition and estimation have been extensively studied using image processing techniques and artificial intelligence (AI). Image processing techniques for crop row estimation generally involve identifying weed density, providing guidance, and extracting overlapping data for region-specific processing, which allows for the estimation and recognition of crop rows [
19]. Curve recognition is typically performed by extracting feature points from images and progressively estimating paths [
20]. However, image-based crop row estimation is often susceptible to misrecognition due to crop growth, requiring multiple pattern settings. Crop row detection, based on machine vision, generally faces challenges such as a low detection accuracy and a suboptimal real-time performance. Furthermore, complex field conditions—such as a high weed density, poor lighting, and the presence of shadows cast by vegetation—pose significant challenges to crop row detection [
21,
22]. Recent advancements in AI-driven image recognition techniques provide solutions to these challenges [
23]. For example, semantic segmentation using deep neural network architectures, such as Fully Convolutional Networks (FCNs) and U-Net, has been successfully applied to tea-picking operations, enabling the extraction of tea row contours for accurate crop row estimation [
24]. Additionally, deep learning models such as R-CNN and SSD have demonstrated efficacy in crop row estimation in rice fields [
25]. In strawberry fields, convolutional neural networks (CNNs) have been utilized to segment RGB images into crop and non-crop areas, effectively handling uneven contours [
26]. For potato crops, U-Net with a VGG16 backbone has been employed to adaptively adjust the visual navigation line position according to crop growth stages, providing accurate navigation line detection [
27]. Lettuce crop row estimation has utilized vegetation indices derived from captured images, combined with the Progressive Sample Consensus (PROSAC) algorithm and distance filtering, to reliably extract crop row centerlines and achieve real-time recognition at 10 frames per second (FPS). The YOLOv8-seg model has been proposed for its balanced performance in both real-time detection and accuracy, out-performing other segmentation models such as Mask R-CNN, YOLOv5-seg, and YOLOv7-seg, with improvements in mean Average Precision (mAP50) of 10.8%, 13.4%, and 1.8%, respectively. Furthermore, it achieved a detection speed of 24.8 FPS on a Jetson Orin Nano standalone device [
28]. In initial tests on onion fields, the YOLOv8-seg model demonstrated its ability to reliably estimate crop rows under challenging conditions, such as a high weed density and variable lighting. Given these findings, the YOLOv8-seg model is considered to be particularly suitable for real-time performance and a high detection accuracy, especially for onion crop row estimation using AI image recognition.
To reduce environmental impacts and improve driving precision, the adoption of electric crawler-type machines has been considered. It is estimated that the electrification of agricultural machinery will reduce agricultural carbon dioxide emissions by 44–70%, significantly contributing to environmental sustainability [
29,
30]. Combustion-engine-driven machines, such as tractors, have slower acceleration and deceleration compared to electric vehicles, making high-precision autonomous driving more challenging. Electrification enables more precise automatic driving [
31]. Among agricultural machinery, both wheeled and crawler types are commonly used. Crawler machines exert a lower ground pressure and offer a greater traction efficiency than wheeled machines, which supports the development of electric crawlers [
32]. Research on the automatic driving of crawler-based systems has shown high levels of driving accuracy, with RTK-GPS and IMU systems achieving lateral and directional accuracies of approximately 1 cm and 0.2°, respectively [
33]. For these reasons, this paper adopts the electric crawler type for its operational benefits.
Therefore, this study aims to automate weeding cultivation tasks by developing an electric vehicle (EV) crawler that replaces traditional tractor machines in agricultural operations. A system is developed to perform autonomous driving by recognizing crop rows through AI-based image recognition and the estimation of crop lines. A comparative analysis is conducted between manual and automatic driving systems, using both tractors and EV crawlers, to evaluate the tracking performance of the developed system. The practical implementation of crawler mechanisms is highly promising; however, challenges such as cost, scalability, and data quality and accessibility arise during operation. Overcoming these challenges will require not only technological advancements, but also the optimization of system design and resource management. In the implementation of AI systems, it is crucial to develop strategies for efficiently utilizing crawlers and aim for sustainable operation. This study is believed to contribute to addressing these issues.
3. Experiment Results
The experiment was conducted in the fields of Yahagi Agriculture Co., Ltd., located in Tsubetsu Town, Abashiri District, Hokkaido in the weeding season of 2024. The following two types of tests were performed: one involved simulating crop rows using green hoses, and the other involved actual weeding with a cultivator in onion fields. A green hose was used as a substitute for crop rows. The rationale for this choice lay in the hose’s ability to easily form straight or curved lines, facilitating the evaluation of tracking accuracy toward the target. Additionally, the hose could be prepared regardless of the season, offering a high versatility in simulation environments. This characteristic enables efficient and consistent experiments that mimic real field conditions. Both manual and automatic operations of the tractor and EV crawler were recorded. The systems were driven at approximately 5 km/h with a weeding cultivator attached, and their tracking performance was compared.
An additional 50 images were incorporated into the training dataset to enable the system to accurately detect the green hoses. The accuracy of crop row tracking was evaluated by recording the lateral errors based on the AI image recognition results.
Figure 13 illustrates the experimental setup using the hoses.
To evaluate the tracking accuracy of the developed autonomous driving system, actual weeding operations were conducted in onion fields using a cultivator. The manual operation of the tractor by the farmer, as well as both the manual and autonomous operation of the EV crawler, were recorded. The accuracy of crop row tracking was assessed by recording the lateral errors from the AI image recognition results, similar to the experiment with the green hoses.
Figure 14 illustrates the experimental environment in which the system followed the actual onion crop rows.
Figure 15 shows the AI recognition results for the green hose during operation.
Figure 16 demonstrates the tracking accuracy of the autonomous driving system. As illustrated in
Figure 16, the system maintained a maximum error within ±2 cm while following the path. These results suggest that autonomous driving using AI image recognition on the EV crawler is feasible.
Figure 17 presents the AI recognition results for the onion crop rows during operation.
Figure 18 illustrates the tracking accuracy of the manually driven tractor, while
Figure 19 shows the tracking accuracy of the manually operated EV crawler.
Figure 20 demonstrates the tracking accuracy of the EV crawler during autonomous operation.
Table 10 lists the standard deviation values for each scenario. Based on the data in
Figure 20 and
Table 10, it is evident that the tracking accuracy of the EV crawler was highest during autonomous operation.
During the experiments, the system operated for a maximum of approximately three hours per day, consuming about 40% of the battery.
4. Discussion
This paper focused on the automation of weeding cultivation and aimed to develop a control system for an EV crawler-type weeding robot. The research encompassed the development of the EV crawler, crop row recognition using AI, the design of an automatic crop row tracking system, and simulated driving experiments on field paths. In terms of crop row recognition with AI, significant progress was made in miniaturizing the image recognition PC. Additionally, the recognition of crop rows in diagonal patterns was successfully achieved.
The robot developed in this study achieved a high level of accuracy (1.4 cm) by combining AI-based image recognition technology with an electric system. The AI technology accurately recognized crop rows within the field and operated efficiently without being affected by environmental conditions (lighting, soil changes, or plant growth patterns). This improvement in accuracy was due to the AI’s ability to dynamically adjust the path by making real-time decisions based on the complex conditions within the field. Additionally, the adoption of the electric system provided a higher responsiveness and precision compared to traditional engine-driven systems. This enables stable operation and will significantly enhance the accuracy of agricultural automation tasks.
In contrast, other agricultural robot systems, such as John Deere’s autonomous tractor or Naio Technologies’ robots (e.g., Oz), rely on GPS technology and basic image recognition, usually operating with an accuracy from around 2 to 5 cm. These systems, being based on GPS and cameras, are particularly sensitive to weather and environmental conditions. For instance, fog, strong sunlight, or soil moisture can impact the accuracy of the sensors, potentially leading to misidentifications. In contrast, AI-based systems can handle these variations more flexibly, achieving highly accurate crop recognition and movement. Furthermore, the electric system experiences less mechanical wear and allows for precise control, enabling the robot to operate steadily and accurately even on complex or uneven terrain—an advantage over other systems.
5. Conclusions
In this study, we developed an EV crawler-type weeding cultivator robot used in an organic onion field. The experiment results demonstrated a superior navigational precision of about 1.4 cm. The automatic navigation precision outperformed human-operated controls. This high accuracy can be attributed to both hardware and system components. The crawler-type design enhanced mobility and stability in rough agricultural fields, while the shift to electric power improved responsiveness and precision. The integration of AI-based image recognition technology further enhanced accuracy, enabling the robot to reliably identify crop rows, even in complex field conditions with varying lighting, soil, and growth patterns.
The robot’s performance highlights the potential of combining AI with electric systems for agricultural automation. The system’s ability to recognize both straight and diagonal crop rows, while minimizing misrecognition errors, allowed it to adjust its path dynamically in real time. This capability ensures efficient and accurate task execution across diverse environments, offering a significant improvement over traditional systems.
Despite these advancements, there are potential challenges regarding the autonomous driving system that need further attention. One potential issue is the AI model’s ability to accurately differentiate between crop rows and other elements in the field, such as green straight lines, which could be mistakenly recognized as crop rows. While we briefly mentioned the possibility of this situation, it remains a concern that could affect the robot’s performance, especially in fields with varying plant growth patterns, shadows, or similar visual cues. Future research should focus on developing learning models that are more robust to such confounding elements, potentially incorporating additional data sources like plant behavior, growth stages, and environmental changes over time.
Additionally, while the current system performed well in the experimental environment, its behavior may vary in different weather conditions and environments. For example, extreme weather such as heavy rain or high winds could impact the robot’s navigational precision or its ability to detect crop rows reliably. To address this, future studies should compare the results of autonomous driving in various weather conditions and field environments. This would help to identify any environmental factors that influence the robot’s performance and could lead to improvements in the model’s adaptability.
Another limitation of the current research is the robot’s handling of large-scale, dense crop fields, where the inter-row movement systems may face challenges. Enhancing the robot’s ability to operate efficiently in such environments, while maintaining its high level of precision, will be a key area of future development. Furthermore, while the integration of AI has improved task execution, the model’s ability to adapt to highly diverse crop behaviors and irregular field layouts needs further refinement.
In terms of future research directions, one approach could be to expand the AI learning models to focus more on crop behavior patterns over time, rather than relying solely on visual recognition. By incorporating a broader set of environmental and biological data, the model could be trained to make more accurate predictions in real time. Additionally, exploring multi-sensor fusion, combining visual, thermal, and possibly even auditory data, could improve the robot’s ability to recognize crops in challenging conditions.
In conclusion, while the robot’s performance demonstrates significant promise, there are several areas for further research and improvement. These include refining the AI model to minimize misrecognition, enhancing its adaptability to diverse environments, and developing more efficient systems for large-scale agricultural tasks. By addressing these limitations, future models can improve their applicability and reliability in various agricultural settings, supporting sustainable farming practices and reducing reliance on manual labor.