Next Article in Journal
Remote Sensing Applications in Coastal Areas
Next Article in Special Issue
Biot-Granier Sensor: A Novel Strategy to Measuring Sap Flow in Trees
Previous Article in Journal
Distributed Target Tracking in Challenging Environments Using Multiple Asynchronous Bearing-Only Sensors
Previous Article in Special Issue
Applying Infrared Thermography to Soil Surface Temperature Monitoring: Case Study of a High-Resolution 48 h Survey in a Vineyard (Anadia, Portugal)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Agricultural Robotics for Field Operations

1
Agricultural University of Athens, Iera Odos 75, 11855 Athens, Greece
2
Fraunhofer IPA, Nobelstr 12, 70569 Stuttgart, Germany
3
Wageningen Plant Research, Wageningen University and Research, P.O. Box 644, 6700 AP Wageningen, The Netherlands
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(9), 2672; https://doi.org/10.3390/s20092672
Submission received: 30 March 2020 / Revised: 24 April 2020 / Accepted: 5 May 2020 / Published: 7 May 2020
(This article belongs to the Special Issue Advanced Sensors in Agriculture)

Abstract

:
Modern agriculture is related to a revolution that occurred in a large group of technologies (e.g., informatics, sensors, navigation) within the last decades. In crop production systems, there are field operations that are quite labour-intensive either due to their complexity or because of the fact that they are connected to sensitive plants/edible product interaction, or because of the repetitiveness they require throughout a crop production cycle. These are the key factors for the development of agricultural robots. In this paper, a systematic review of the literature has been conducted on research and commercial agricultural robotics used in crop field operations. This study underlined that the most explored robotic systems were related to harvesting and weeding, while the less studied were the disease detection and seeding robots. The optimization and further development of agricultural robotics are vital, and should be evolved by producing faster processing algorithms, better communication between the robotic platforms and the implements, and advanced sensing systems.

1. Introduction

Major technological advancements in agriculture have drastically transformed several processes, both in crop and livestock production systems, during the last decades. These advancements are mainly related to the minimization of operational and production costs, reduction of environmental impact and optimization of the overall production cycle. Focusing on crop production, a series of optimization models and software tools have been developed so far on field-operation level. This progress, in parallel with the technological advancements and equipment in field machinery, has provided radical solutions to several challenges that modern farmers face. In crop production systems, one of the most significant issues is connected to human labour-intensive operations. These are, mainly, field tasks (such as sensitive fruits harvesting and intra-row weed control) that are more difficult to be executed by traditional field machinery, and human workers are employed. This has brought the increased need for autonomous tractors and robotic platforms to be used in the crop field operations, currently developed at research stage [1].
Field operations in agriculture are quite complex, and various issues should be addressed to allow an effective transition towards the robotics era. To build a robotic solution, an overall system analysis of the field operation should be conducted, together with a cost–benefit analysis [2,3]. Such a system should comply with very specific requirements, such as lightweight, small size, autonomy, intelligence, communication, safety, and adaptability, to execute the potential task effectively [4]. The relatively smaller size of autonomous machines, compared to conventional tractors and implements, contributes to the reduction of the soil-related issues and in particular soil erosion and soil compaction, that was provoked by large and heavy modern farm machinery [5]. To automate these field operations, a decomposition of these tasks should be performed to transform them into discrete robotic functionalities. This can be achieved by classifying the agricultural tasks into deterministic (tasks that can be designed and optimized, in advance) and reactive (tasks that are associated with behaviors that should deal with unexpected conditions) [6].
The main challenges that agricultural robots face are associated with both universal and task-specific issues. Universal issues are related to terrain assessment [7,8], route planning [9,10], safety issues, especially focusing on human detection [11], and fleet of robots [11,12,13]. Task-specific issues are related to specifications related to crop architecture, crop or pests detection and classification, and precise application of inputs. Most of these issues are related to the vision system, the robotic actuation system, the navigation system in semi-structured agricultural environments, and the intelligence to control both the robotic platform and the implement.
There have been several reviews focusing on agricultural robots for field operations based on specific tasks (such as agricultural navigation, transplanting/seeding, pruning, weed control, harvesting, traceability and multi-robot interactions) and systems assessment [14]. One study reviewed a group of field operations applied only in open arable farming systems [15], while the digital perspective of mainly commercial agricultural robots has been reviewed as well [16]. Finally, a short review based on the methodological approach, the field operation and the farming environment has been presented, under the demonstration of an emulation tool for agri-robotics [17]. As weed control robots are among the most developed agricultural robots so far, a review has been conducted regarding only autonomous weed control systems [18], while there has also been a crop-specific literature review about robotic systems applied to strawberries [19]. Even though the scientific contribution of these reviews is highly important, the contribution of this study is to provide a holistic review of agricultural robotic systems that execute all the main field operations, for all cropping systems and in both open-field and greenhouse environments.
The aim of this paper was to study the specific issues related to main field operations in crop production. More specifically, the main challenges leading to the concept of the current review are: (i) to identify and emphasize the currently existing robotic systems developed for each field operation, regarding different crop production systems; (ii) to provoke further development and optimization of existing agricultural robotic systems in general and, more specifically, the development of smarter perception and actuation systems; and (iii) to propose the key challenges and directions for future agricultural robotic developments.
This review paper covers both scientific and commercial agricultural robotic systems applied in various field operations (such as weeding, seeding, disease detection, crop scouting, spraying and harvesting) for different farming environments (arable, greenhouses, vineyards and orchards). The structure of the present work is as follows: the followed framework in this review is showcased in Section 2. In Section 3, the operational classification and analysis are both presented and allocated in the main field operations sub-sections. Finally, the overall discussion of the review is shown in Section 4, followed by Section 5 where general conclusions are underlined.

2. Review Framework

In the pre-processing stage of this review, a series of theoretical considerations were taken into account. Throughout the analysis, a series of steps were established, as presented in Figure 1: (i) development of the review protocol regarding the eligibility criteria; (ii) search for research studies and commercial resources, and selection of the ones satisfying the eligibility criteria; (iii) definition of the classification framework to be applied in the literature review, in order to classify the material and build the structure; (iv) selection of studies to be included in each category of the classification framework; (v) analysis of the selected studies; and (vi) representation of results by comparing studies.
Regarding the eligibility criteria, the studies that are related to robotic systems applied in crop production systems have been included as a wide boundary of this study. More specifically, the included references should be related to robotic systems that conduct specific field operation(s) in crop production systems in agriculture. The referenced robotic systems may have been developed in various environments (i.e., open-field, greenhouses, laboratory, etc.). Hence, they should be applicable to any in-field operation, excluding out-of-field transportation operations.
In addition, a number of literature-related eligibility criteria are set to be followed. Those are: (1) the work should be published in the English language; (2) the included studies should be research articles published in scientific journals, conference proceedings, or other sources (commercial websites, research projects, etc.); (3) they should have been published within the last 20 years (without including the current year), as any technology used prior to this is probably obsolete today. Publications in journals that are not research articles (such as reviews) are out of the scope of this review and they are excluded.
The included studies have been withdrawn from open online sources (such as open-access journals, websites, conference proceedings, etc.) and by a systematic literature review in various electronic depositories, namely Web of Science, ScienceDirect, Scopus and SpringerLink. The primary literature research was conducted at the end of 2019. Thus, any newly emerging publication will have been possibly missed. Due to the nature of the reviewed field, grey literature (technical/research reports, doctoral dissertations, etc.) have been knowingly permitted by the authors in order to cover as much as possible in this review.

3. Operational Classification and Analysis

The existing literature of robotic systems in agricultural field operations was classified according to major field operations, as follows: (1) weeding, (2) seeding, (3) disease and insect detection, (4) crop scouting (plant monitoring and phenotyping), (5) spraying, (6) harvesting, (7) plant management robots, and (8) multi-purpose robotic systems [4]. Given that the current literature is rather wide, a brief summary of the most important literature, based on scientific and commercial resources, has been allocated to these categories and presented below.

3.1. Weeding Robotic Systems

Weeding is one of the most repetitive, tedious and time-consuming activities within the crop production cycle. Especially for developing countries and smallholder farms, weed management accounts for more than 40% of the labor effort [20]. Besides being one of the most difficult tasks, it is one of the most costly, including among other factors significant labor cost, especially in crops that are highly labor-demanding. Moreover, it is a field operation that has a substantial negative impact on the environment, as large quantities of herbicides are used to keep the field weed-free, to promote plant health and yield. As an example, global estimated loss potential in rice, wheat and maize indicates that weeds account for 46.2–61.5% of potential yield losses, and 27.3–33.7% of actual yield losses caused by all pests together [21]. For all the above-mentioned reasons, significant attention has been given to weeding robots both by private companies and academia. A number of such weed robots have been presented in a short review [22], while in the current study a wider review has been conducted regarding the major field operations, including weeding robots.
Traditionally, there have been two approaches for weeding; namely, mechanical weeding, and chemical weeding. Mechanical weeding describes the task of destroying weed plants, mainly by plucking them out, burning them or cutting them. For each of the three above-mentioned methods of mechanical weeding, a variety of tools are available. In order to remove the weeds by mechanical means, the system should first detect the row that it will work on, and then cut or pluck out the weeds. The precision in row detection recorded with mechanical weeding robots was less than 25 mm [23,24], followed by other systems with precision less than 6 mm, [25,26] or higher, up to 3 cm [27,28]. In addition, the performance of the weed removal (expressed also as weeding efficiency) has been evaluated in systems that presented rates higher than 90% [24,29], and 65% and 82%—for two different presented methods [30]. Apart from those, there are a series of mainly commercial mechanical weeding robots that have no performance metrics provided [31,32,33,34].
On the other hand, chemical weeding refers to spraying the weeds with herbicides that have a toxic effect, thus eliminating the weeds, and it has been popular because it requires considerably less labor. Herbicides manufacturers have also developed active ingredients that are only toxic to weeds and not crops, allowing for extensive spraying over the entire field. However, a number of chemical substances, especially in the EU, have already been restricted in their use, or are going to be restricted in the near future, due to the negative impact on the environment and human health. In addition, weeds have already developed chemical resistances, making them harder to eradicate. The main solution proposed to reduce the pollution caused by herbicides and delay the appearance of resistant weeds is selective/spot spraying, spraying only the weed and not the entire field. This would require even more working hours to be spent on weeding. On the bright side, recent developments in computer vision technologies have allowed manufacturers to automate that process, and integrate high-level software into spraying/weeding robots.
Similar to mechanical weeding robots, the effectiveness of the system for spotting weeds is among the most critical performance metrics for the chemical weeding robots. By using a Drop on Demand (DoD) system—a system that detects the weeds within the plant row and selectively shoots droplets of herbicide onto those weed leaves—100% effectiveness was achieved, which is the highest among chemical robots [22,35]. In parallel, 98% and 89% detection accuracy were presented in a double experiment of a different vision-based weeding robot [36]. Finally, effectiveness (weed detection and destruction) of more than 85% under ideal conditions is underlined by the EcoRobotix weeding robot [37]. While a series of commercial chemical weeding robotic systems are available, almost no performance metrics exist [38,39,40,41]. The perception sensors used for weeding offer similar levels of performance. All robots make use of a camera, where RGB (Red–Green–Blue), infrared (IR) and web cameras are the most commonly used. There also exist sensors that allow the robot to identify its surroundings and its relative position to the weeds, using acoustic distance sensors, gyroscopes, laser range finders, Inertial Measurement Unit (IMU), etc. Table 1 outlines the most significant chemical and mechanical weeding robotic systems with their main characteristics in terms of perception sensors, level of weed detection, control method and performance rates (where they exist).
To conclude, weeding robots have already been under development over the last years, and there are also a number of commercial robots on the market. Most of the major problems related to the existing weeding robots are being dealt with—mainly, the use of recent advancements in computer science (e.g., deep learning) and the integration of more accurate navigation systems (e.g., RTK GPS). What remains to be done is to improve their performance, increase working area and working speed, and increase the accurate detection of weeds for selective spot spraying. There are a few key issues that should be taken into account; i.e., how the existing robotic systems could perform on crops with similar operational requirements, to expand their potential use and the overall development of weeding robots in crops with low competitiveness towards weeds, such as specific vegetables. In most weeding robots, the vision sensors are currently used in combination with distance identification sensors. This, though, could be potentially enhanced by even higher resolution weed detection sensors for selective spot spraying. Generally, existing robots focus on above-ground weed detection. Possible solutions for future weeding robots may include the detection of weeds in their primary growth, when the weed sprouts are still below the ground. Such systems could be equipped with sensors measuring, for instance, the soil electric conductivity.

3.2. Seeding

Seeding is one of the fundamental tasks in a crop production cycle and contributes significantly to the labor cost. Seeding machinery has evolved to offer precision seeding solutions. Since recent publications have shown the importance of plant densities to a high yield [43], precision seeding has gained more interest in the scientific community, which as a result has led to the development of several seeding robots in recent years (Table 2).
The main performance metrics presented in the existing literature are connected either to accuracy rates (%), regarding the number of seeds that have been accurately placed, or to position errors. The main targeted crops for the existing reported seeding robots are mainly cereals, with high accuracy rates reported regarding wheat and rice crops [44,45]. On the other hand, an autonomous seeding robot more suitable for wet-fields has been proposed, but with no performance metrics [46]. Finally, there has been progress in commercial seeding robots made by Fendt [47,48], which can be used as an intelligent fleet of seeding robots or individual units. The most important information for seeding is the depth and the position that the seed is going to be planted. As a result, force, pressure, angle and displacement sensors are used in combination with wheel encoders, compasses and displacement sensors.
In conclusion, the last years’ seeding robots have been developed for seeding both at com0mercial and research level, but to a limited number due to the complexity of the operation. Even though the current advancements in sensors play a crucial role in facing the issues related to these robots, they should be optimized further by minimizing the errors and maximizing the accuracy rates. There are a few key challenges identified for seeding robots, such as the working speed of the robot, as it is one of the main factors for precision seeding, making its calibration crucial for each robot. Moreover, the need to improve speed and accuracy, with better vision systems. Better image detection systems that use more advanced algorithms are already under development [50]. Before seeding, such vision systems should be able to identify the soil surface quality and structure, the existence of any weeds, and possible obstacles that may occur. After seeding, the seeding performance should be evaluated, especially in fields with an uneven surface.

3.3. Disease and Insect Detection

Robotic application in disease and insect management (detection and control) has very recently gained attention, due to its complexity and the lack of accurate and efficient conventional systems. Disease detection is an important part of the production cycle, as diseases can cause significant economic damage if not detected during the early stage. Insect detection has not gained a lot of attention, as many insects are located underneath the leaf, inside the buds or even underground, which makes it extremely difficult to accurately detect them and, as a result, monitor them. In Table 3, the main perception features and the highest detection accuracy for each cited work are showcased for a number of crops and diseases.
Disease detection is mainly a visual task, and therefore all robots incorporate a vision-based system. Detection accuracy is defined as the ratio of the identified diseased plants to the total number of the diseased plants. All the reviewed robots use a color camera, which is low-cost and easy to use, with the second most used camera type being a multispectral camera, which is moderately expensive and requires complex higher computational power. In principle, all system configurations with multi-/hyper-spectral sensors were combined with color cameras. At this point, it is worth mentioning that using the more complex system does not always yield the best results. For the detection of powdery mildew and tomato spotted wilt virus (TSWV) in greenhouse pepper plants, using a color camera showed high accuracy, while the use of a multispectral camera achieved an accuracy of 80% and 61% for these two diseases, respectively [51,52]. A different approach was introduced for the detection of Xylela fastidiosa in olive trees, fusing a series of sensors [53].
Simpler disease detection robots are based mainly on RGB image inputs. With regards to this, powdery mildew leaf disease has been detected in strawberry plants by a machine vision system, that was based on artificial cloud lighting conditions, and achieved accuracy from 72–95% under two different lighting conditions [54]. Similarly, RGB cameras have been used for Pyralidae insect detection on tomato and rice plants with accuracy of about 94% [55]. Moreover, eAGROBOT was used to detect a series of diseases, and presented about 90% accuracy for both cotton plants and groundnut plants [56].
To sum up, disease detection robot technology is in its infant stages. From what is available at this point, there are three major problems related to disease detection in robots: (i) the lack of image databases for each disease in order to train the detection/classification models; (ii) the slow image processing, especially when using large volumes of images such as hyperspectral; and (iii) the non-uniform lighting conditions in the field. The lack of available image datasets is gradually being addressed, as open-access agricultural databases are becoming available [57] and new data synthesis methods are presented [58]. At the same time, novel lighting systems are also being proposed for the detection of various diseases such as powdery mildew [54]. As for insect detection, it has not gained attention due to the constraints related to insect features (such as size, color, shape, etc.). Potential future vision systems could detect insects by using multiple ultra high definition cameras, that could identify insects covered among the leaves, underneath or elsewhere on the plant. Even in this case, such an insect detection robot should include high-performance algorithms, given the high volume and size of images, in order to perform accurately.

3.4. Crop Scouting

3.4.1. Plant Vigor Monitoring

Plant monitoring is the oldest practice growers have been performing to ensure high yield and quality. Monitoring plant stress so far has been done by searching for visible symptoms, such as change of color, withering, spots, or any abnormality on the plant leaves or the plant in general. However, as remote sensing technology is advancing, new sensors are becoming available for the grower to monitor plant health and stress, even when it is not visible to the human eye. Nowadays, thermal cameras and infrared thermometers are used to measure canopy temperature [59]. Further, various vegetation indexes, such as the Ratio Vegetation Index (RVI), the Normalized Difference Vegetation Index (NDVI) [60], etc., which provide useful information about plant health, photosynthetic ability and more, are used by combining wavelengths. Moreover, expensive hyperspectral sensors are starting to find applications in plant monitoring systems [61].
Various plant monitoring robots have been developed so far, even though they are in their early maturity level, due to the high cost of the sensors that are utilized. As a result, the main focus has been on high-value crops and how robotic systems can effectively replace humans in monitoring tasks. In Table 4, various plant monitoring robots are presented.
The majority of crop monitoring robots have been developed specifically for orchard/tree crops or vineyards. A crop monitoring robot has been experimentally evaluated in laboratory conditions [62] and semi-structured environments [63], and proved to be highly reliable for the estimation of orchard canopy volume and vigor, by using a series of sensors. Crop monitoring robots for vineyards have been widely developed, even though there are no clear quantified performance metrics. Most of them are commercial crop monitoring robots (such as [64]), or robots that have been developed under EU-funded projects ([65,66,67]). In all cases, a series of vision-related sensors are included, such as various types of cameras (RGB, hyperspectral, multispectral, etc.), and LiDAR sensors are incorporated on the robots, in order to monitor volume, vegetation, and health of the plants. Given that vineyards are usually cultivated in hilly or mountainous terrains, a vineyard monitoring robot deals with a series of challenges related to terrain traversing, mapping, and localization [68].
Similar perception systems have been implemented in crop monitoring robots for crops such as Canola, which presented maximum measurement error up to 2.5% [59]. In this specific case, the platform is mounted on a swather. In addition, Earthsense has commercially developed the Terrasentia crop monitoring robot, suitable for various crops such as corn, soybean, wheat, etc.[69].
More generic crop monitoring robots handle agricultural operations, such as soil sampling [70]. CO2 concentration levels may have a significant effect on crop growth. Under this scope, a hexapod robot has been developed for the collection of agronomical information, such as soil nutrients and plant growth. The monitored CO2 concentration levels were up to 2500 ppm [71].
To conclude, plant vigor monitoring robots face challenges such as illumination, background separation, and data manipulation. Various lighting conditions due to weather conditions have been faced so far, by artificial means and by following strict timetables on image collection processing. Background separation is due to neighbor plants that create occlusions in the visual sensors. This problem can be faced either by better algorithms, that could capture these visual interpolations, or by using multiple cameras. The huge amount of data collected is an issue that could be manipulated by faster processing algorithms and more performant hardware. In the future, plant vigor monitoring robots could include smarter vision systems, i.e., high definition cameras that will automatically correct low-light images and apply artificial light where needed, in order to produce uniform image data.

3.4.2. Phenotyping

Parallel to the development of plant monitoring robots, phenotyping robots have been under development. Plant monitoring and plant phenotyping have a lot in common, as in both cases, vision systems are used to gather as much information as possible about the plant/fruit under inspection. The difference is that phenotyping robots focus more on morphology, plant growth and traits that contribute to crop performance, while plant monitoring robots focus on the state, health and chemical composition of the plant. Numerous sensors used in phenotyping robots have been reviewed in [72]. The developed phenotyping prototypes are presented in Table 5, for various levels of autonomy.
Fully autonomous phenotyping robots have been developed in sorghum plant field trials, presenting stalk detection up to 96% [73]. These robots also measure light penetration, leaf erectness, leaf necrosis and GRVI (Green-Red Vegetation Index). Similar autonomous phenotyping robots in sugar beet crop (Mobile and Bettybot) analyze in detail the geometric and colorimetric evolution of the plants [74]. Bonirob, an autonomous phenotyping robot, uses multisensor data fusion to measure population density, plant distribution, plant height, stem thickness, etc. [75].
In addition to the autonomous robots, a fixed-site, fully automated phenotyping platform has been applied in rice, maize and wheat plants [76]. This platform may provide highly accurate measurements on canopy growth throughout crop lifecycles, including plant height root mean square (RMS) error up to about 1.9 cm. In sorghum crop, a tractor-mounted autonomous robot resulted in highly repeatable measurements (such as plant height, width, convex hull volume and plant surface area) that are quite close to the in-field manual measurements [77].
Not all of the phenotyping robots are fully automated. Two phenotyping robots are able to work collaboratively in cotton crop, as an autonomous ground vehicle for individual plant data collection, and as a mobile observation tower that inspects the entire field [78], including RMS (Root mean square) errors for plant height less than 0.5 cm, RGB to IR image calibrations of 2.5 pixels, and less than 1 °C for temperature. Finally, another semi-autonomous phenotyping robot has been tested in energy sorghum crop [79]. The average absolute error achieved by the algorithms used in this robot was 13% for stem width estimation, and 15% for plant height estimation.
In conclusion, plant vigor monitoring and plant phenotyping are two similar tasks, and therefore face identical challenges. Illumination is a major problem, as lighting conditions are constantly changing. So far, the way researchers have dealt with this problem is by using artificial lighting to ensure uniform illumination, and by collecting their data at the same time of the day [77]. Another challenge is the separation of the background, mainly due to occlusions caused by neighbor plants. Various solutions for the above-mentioned problem have been suggested. Some focus on the crop, for example, cultivating and using wider rows, but the major focus is on the technology used, and mainly on improving the algorithms [80] used, as this does not require changes in cultivation practice. The final challenge is to deal with the amount of data generated. Datasets have to be efficiently processed, the workflow has to be automated, especially when data from different types of sensors are gathered, and pipelines that translate data to the consumer must be built. Computer scientists are already introducing better and faster image processing algorithms, and pipelines that could find use in agricultural applications [81,82].

3.5. Spraying

Besides weed management, where herbicide sprayers are used, spraying robots can be used for pests (diseases and insects) and also for liquid fertilizers (foliar). As a result, the farmer is exposed to large quantities of those toxic active ingredients no matter the protective measures that are taken. It is, therefore, crucial to introduce spraying robots to prevent possible health hazards. Due to the rapid development of computer vision and artificial intelligence, robotic sprayers feature novel intelligence systems that enable selective spraying, compared to conventional uniform spraying across the crop. Robots could make use of such technologies, reducing agriculture’s environmental impact as well as consumer exposure to pesticides, and preventing the development of resistance to those substances by the targeted organisms. An overview of the spraying robots can be found in Table 6.
Various approaches have been proposed when designing a spraying robot, mainly due to restrictions and possibilities of the operating environment. In greenhouses, two approaches have been identified: robots that move using the greenhouse piping system [83,84], and robots that navigate between rows without being mounted on those pipes [85,86,87]. Moreover, the ability to spray selectively requires an accurate detection system, and therefore advanced sensors need to be mounted on the robot [88,89]. Finally, the spraying system (nozzles) could be mounted directly on the spraying platform [83,84,85,86,87,90], or it could be mounted on a robotic manipulator with various degrees of freedom (DOF). The manipulator used by [91] had 3 DOF, the one used by [89] 6, while the one from [88] had 9 DOF.
Based on the provided results, the execution time of the spraying operation is presented [86,89], while in most cases the effectiveness of the system is reported. The effectiveness of the systems is demonstrated by providing the successful spraying rates, the error of the system [85,87], or both performance metrics [83,88,90]. Finally, only a few of the systems consider real-time detection of the potential spots to be sprayed [88,89,90], while the majority of them do not include real-time detection [83,85,86,87,91].
It can be concluded from the metrics provided in the reviewed papers that spraying robots perform reasonably well, with non-selective sprayers achieving coverage of about 92% and selective sprayers being able to identify and spray 85% of the diseased areas [83]. The speed, however, of selective sprayers is low, with the total spraying time for every seedling in a hectare cultivated with vegetable crops being higher than 30.5 hours [89]. Although, it has to be mentioned that the time needed can be considerably less since not all plants need to be sprayed. Some of the solutions suggested to decrease the time needed for spraying involve using more spray nozzles or more than one platform, and increasing the speed of the detection algorithms. All these are directly connected to the further development of detection sensors, in order to reach even better accuracy rates in the future.

3.6. Harvesting Robotic Systems

Harvesting is one of the most labor-intensive and repetitive tasks, while it is part of all production cycles in agriculture. As a result, several robotic systems have been developed by private companies and universities/research centers to automate this procedure.
There are two types of robotic harvesters: bulk (every fruit/vegetable is harvested) and selective (only the ripe/ready to be harvested fruit are collected). In this review, the focus is mainly on selective robotic harvesters, as this has attracted the most attention from research development. The two major performance metrics of selective robotic harvesting are picking speed and picking rate (i.e., the number of fruits successfully picked out of the number of ready to be harvested fruit [92]). Harvesting must be performed within a certain time slot when the crop is mature, and the majority of the crop has to be harvested and all the above executed without damaging the crop and the plant.
The majority of the harvesting robots focus on strawberries, a high-value crop, that suffers from a high production cost mainly due to labor cost, particularly during harvesting [93,94]. For strawberries, the fastest picking robots were the strawberry harvesting robots in [92,95], with picking speeds of 7.5 and 8.6 seconds per strawberry, respectively, and the Berry 5 robot from Harvest Croo, with a claimed picking speed of 8 seconds per fruit [96]. Other strawberry harvesting robots achieved picking speed up to 10 s/fruit [97] and 11.5 s/fruit [98]. However, having a high picking speed is not enough, as a high picking rate is also important. Having unharvested fruit means that the grower will have to manually harvest these fruits, which is not additional labor compared to manual harvesting, but it is still labor which adds to the total labor cost. The highest picking rate presented was 86% [64]. A range in picking success rates, from 0% up to 64%, has been shown for various categories to evaluate robots’ harvesting performance [99]. It is worth mentioning that for strawberry harvesting, it appears to be a trade-off between picking speed and picking rate, as fast picking speeds are accompanied by poor picking rates and vice versa. Apart from the abovementioned robots, there is, also, a wide variety of commercial strawberry harvesting robots with no performance metrics provided. Dogtooth technologies developed a selective picking robot for table-top growing systems [100], Agrobot developed the E-series, an autonomous selective strawberry harvester with 24 robotic arms [101], while Octinion presented Rubion harvesting robot [102].
Other crops that have gained attention are apples and tomatoes. The apple fruit has attributes that make it easy to harvest robotically. Red varieties are easy to distinguish among the canopy, while the hard nature of the fruit makes it easier for grippers to harvest it without damaging it. The fastest apple harvesting robots have reported picking speeds of 7.5 s per apple [103] and 9 s per apple [104], while the highest picking rates showcased were 89% [105] and about 90% in high-density apple orchards [106]. Tomatoes are cultivated worldwide, and in terms of vegetable crop production, the tomato only falls behind the potato. The fastest tomato harvester had a picking speed of 23 s per tomato [107], while the most reliable one had a picking rate of 87% [108]. In cherry tomatoes, the picking speed was 8 s per tomato bunch without including the moving time [109]. Metomotion [110] and Root-AI [111] have their own selective commercial tomato harvesters, as well.
Similar to apple robotic harvesters, other references regarding various fruits’ robotic harvesters are included in this study. Citrus harvesters, such as the commercial one by Energid, manage to collect each orange in 2–3 s [112]. Another fruit that has gained attention is cherries, which are also a fruit that can be easily detected by robotic harvesters, with picking speeds up to 14 s per fruit having been shown in cherry orchards [113]. Moreover, Agribot, a manually driven fruit robotic harvester, needs two seconds to grasp and detach each fruit [114].
Vegetables, in general, represent a promising group of crops in which robotic systems could be implemented. Considering this, a series of different robotic harvesters have been developed. A cucumber robot claimed picking rate of 80%, while its average picking speed was 45 s per cucumber [115]. For eggplants, the picking rate was about 62%, while it needed 64 seconds to harvest an eggplant [116]. The Sweeper robot, that was an EU-funded project, is a selective sweet peppers harvesting robot that, achieved an average picking speed of 24 s/pepper, based on the project’s preliminary results [117]. In asparagus crop, the achieved picking speed was 11.9 s per asparagus, when the robot harvested two asparagus at the same position, while in the case that there was only one asparagus to be harvested in each position, the picking speed was 13.7 s/asparagus [118]. Sparter, an asparagus harvesting robot, is also commercially available by Cerescon [119].
Apart from the aforementioned vegetables and fruits, there are certain vegetables that require different harvesting principles due to the challenging features they have, e.g., size and weight. As a result, kinematic analysis and simulation have attracted interest, in order to allow the robotic harvesting of heavy agricultural products, such as pumpkin and cabbage [120]. Regarding fruits, melon and watermelon are representative with similar characteristics. For a melon robotic harvester, the picking rate was about 85% [121], while Stork, that was developed for watermelon harvesting, achieved a picking rate up to 67% [122].
Regardless of the aforementioned categories of robotic harvesters, certain systems refer to miscellaneous crops. Demeter system, designed for the harvesting of alfalfa and sudan (Sorghum vulgare sudanese), achieved harvesting speed up to 2 ha·h−1 [123]. The performance of a mushroom (Agaricus bisporus) robotic harvester regarding its picking success rate was found to be about 70% [124]. In this case, a requirement was to keep scrap rates (i.e., mushrooms that have been damaged in a way that does not meet quality standards) between 5% and 10% (human performance), however, the scrap rate of the robot was about 22%.
A summary of the main reviewed harvesting robots, as well as the highest picking speed and picking rate per crop, are given in Table 7. A direct comparison between robots designed for different crops is irrelevant, as the requirements for each robot are vastly different.
From the reviewed harvesting robots, two were tractor-mounted—the Sparter asparagus harvester from Cerescon and the robotic apple harvester [104]—one was manually driven—the strawberry harvester [95] —and the rest were autonomous. Regarding picking mechanisms, there are two main systems—grippers and suction devices. Grippers are devices that consist of mostly rigid joints and links, and are used to pick up and hold an object, or to exert forces on an object causing rotation with respect to the manipulator reference frame [125]. Suction devices use vacuum to singulate fruit/vegetable, and pull, hold or twist it. For detaching the fruits various approaches have been explored, such as removing the fruit/vegetable by cutting the peduncle using the gripper’s fingers [101], cutting the peduncle using blades mounted on the fingers [92,118] and using the gripper or vacuum suction tool to pluck off the fruit [107,126]. There was one solution, though, from Energid industries, focused on citrus fruit, which did not use any grabbing tools. Instead, it made use of extending rods which aimed at the stem, cutting it and causing the citrus to detach [112]. Prior to the fruit harvesting, it first needs to be localized/identified. The main sensors used for the localization of the fruit were RGB cameras, which in some cases were combined with time of flight sensors [117], infrared sensors [95,101] or laser sensors [109]. Cerescon with its Sparter robot had a completely different approach for localizing the fruit, due to the particular cultivation technique of white asparagus (the asparagus grows in sand beds under the soil) [119]. As a result, proximity sensors were used instead of cameras. Finally, the manipulator’s degrees of freedom ranged from 2 to 7 depending on the crop and the adopted solution, with 5 or 6 of freedom being the most common solution.
A global illustration of the overall performance of each reviewed robot, where the performance metrics are presented, can be found in Figure 2 (multiple robots for each crop). The figure is not provided for a direct comparison but to get a better understanding of how variation in shape, size and fruit characteristics, such as firmness, affect the picking rate and picking speed.
To conclude, harvesting robots cannot directly compete with human labor. Harvesting robotic systems do not perform as intended due to two major problems: localization of the object to be harvested because leaves, fruits and other parts of the plant obstruct the view of the robot; and damage to the fruit due to the sensitive nature of agricultural products. In the future, the performance of harvesting robots should be comparable to those of human workers. In order to do so, they should be able to deal with clusters of fruit and occlusions (better software), and harvest the product reliably without damaging it (hardware). Faster and more accurate vision algorithms should be developed, that work together with soft, conflict-free end effectors, as well as manipulators suitable for trees. Already, various papers are trying to address the above-mentioned problems, for example, Cornell University recently published a paper presenting a deep learning method that improves the performance of a robotic arm end-effector in unstructured and occluded environments [127], and at the same time there is research ongoing for novel end-effector designs that fit the needs of specific crops [128]. At this point it is worth mentioning that it is essential for developers to work closely with horticulturists to come up with suitable tree training methods, that facilitate robotic harvesting and at the same time provide high yield and high-quality products. Finally, the picking speed which appears low could be overcome if the robots are able to work for an extensive period of time. Their overall performance can also be solved if human workers harvests the ends of the field and robots the inner parts, to increase their speed.

3.7. Plant Management Robots

In addition to the main above-mentioned field operations, robots that focus on plant architecture have been included. More specifically, they focus on material handling, pruning/thinning, and string twining. Here, the robots that have been developed for these types of field tasks are presented (Table 8).
Material handling is not a task of high importance in the production cycle, however it is crucial in growing operations such as plant nurseries. Therefore, Harvest Automation Inc. developed an electric, fully autonomous material handling robot for unstructured environments (e.g., greenhouses, hoop houses, etc.) [129].
As has already been indicated, high-value crops get most of the attention, and one of the most studied crops is grapevines. Subsequently, both pruning robots under review focus on grapevines. The first one was developed by a commercial company titled Vision robotics [130], and therefore no technical information is available, while the second one was developed by the University of Canterbury. In order to effectively prune grapevines, the robot would completely cover the plant to block the sunlight, and then use a six-jointed robot arm mounted with a cutter, and three-color cameras in combination with LED lights [131]. Similarly, a robotic end-effector for the pruning of apple trees resulted in the branch cutting force being in a logarithmic relationship with the branch diameter [132].
String twining is a labor-intense task in crops such as high trellis hop, and there is a high demand for mechanization of this operation. To this end, a string twining robot has been developed, which achieved a success rate up to 97% while was moving with speed of 0.19 m s−1 [133].
In conclusion, due to the limited number of publications for each task included in this section, the challenges that have to be dealt with in the future are not clear and solid. However, pruning robots should be able to deal with specific key challenges related to crop architecture, such as height and branches diameter. In this light, suitable sensors should be available on the platform; i.e., force sensors, a vision system, a multi-degrees of freedom end-effector, and, in the background, algorithms that will decide the optimal pruning per plant. Finally, the key challenges that string twining robots face are connected mostly to their performance (faster operational speed and success rate), which is directly connected with the vision system, the end-effector, and the overall platform performance. Overall, there should be more experimental and research studies in the future regarding plant management and architecture robots, in order to optimize their sensing parts and software. In this light, in the current systems, faster execution time cycles could be achieved by adapting higher precision vision systems and faster algorithms.

3.8. Multi-Purpose Robotic Systems

Within the crop production cycle, most field operations are executed over short periods of time and only once or twice per year. Additionally, those time spans are usually the same for all growers. As a result, buying and sharing a robot is not a feasible option. One of the widely accepted solutions, to reduce the depreciation time of machinery and make it affordable to growers, is to add value to the robotic system mostly by designing it for more than one field operation. This approach has been adopted by universities, research centers and commercial companies. In order to allow the robot to execute more than one task, three main approaches have been identified: designing a robot that can be equipped with different implements [134], designing a modular robot [135], or equipping the robot with all the tools from the beginning [136]. It is worth mentioning that all robots are autonomous, however, two of them allow the user to manually control them as well if needed [137,138], and two of them are fixed in place with teleoperation systems being used [139,140]. A list of the most significant reviewed multi-purpose robots and the tasks that they can perform is given in Table 9.
Different design approaches lead to robots that are equipped with a variety of perception and measurement sensors. There are robots equipped only with humidity and temperature sensors [136], robots that combine more sensors (humidity, temperature sensors and color cameras) [139] and robots that are just equipped with color cameras [134]. Finally, as mentioned before, some of the robots are modular, meaning the sensors that they carry change according to the modules used [141], while other robots can be fitted with additional sensors according to the growers’ needs [134].
Most of the multi-purpose robots presented above do not include specific performance metrics. A smart robotic machine equipped with a computer vision system for weeding and variable-rate irrigation schemes presented up to 90% weeding rate and 75% wet distribution area [142]. Similarly to weed detection, fungal disease detection, combined with treatment operations, presented significant fungicide savings (85%) when automatic detection was used [143]. Furthermore, more than 97.7% of maize seedlings were accurately detected, in order to prepare maize fields for tillage operations [144].
To conclude, multipurpose robots perform more than one field operation of the production cycle, and at this point in time, no solutions have achieved the level of modularity and good performance needed, as being able to perform multiple tasks adds complexity to hardware and software. On the bright side, the same solutions that can be used for the individual task robots can also be applied to the multipurpose ones. Besides, for those problems and solutions that have already been mentioned in previous chapters, multipurpose robots have to perform adequately/accurately for each and every task they are developed for. In addition, the ability to fulfill more than one task increases the cost of the robot, due to complex configurations and the variety of sensors needed. A way to justify higher costs is by extending their functionalities using relatively cheap sensors to measure additional parameters.

4. Discussion

The reviewed literature of robotic systems in agricultural field operations was classified according to major field operations; weeding, seeding, disease and insect detection, crop scouting, spraying, harvesting, plant management, and multi-purpose robots. An indicative representation of each category robot is depicted in Figure 3.
The number of references included in this review was 153 in total. They have been allocated, according to the referenced sources, to four main categories, namely research journal sources, conference proceedings, commercial and other sources (such as EU-projects, book chapters, etc). The research journal studies were retrieved from various journals, such as “Computers and Electronics in Agriculture”, “Biosystems Engineering”, “Precision Agriculture”, “Journal of Robotics”, “International Journal of Agricultural and Biological Engineering”, etc. In addition to this, part of these references have been published in conference proceedings, while others are related to commercial robots and sources such as international research projects.
Agricultural robots have been recognized as one of the most promising solutions for solving the challenges agriculture is currently facing. Agricultural robots have been developed for every operation in the crop production cycle, starting with the establishment of a crop by seeding, and ending with the harvesting of the final agricultural product. However, the tasks that have attracted most of the attention are the most labor-intensive ones, namely harvesting and weeding. As most of the agricultural tasks are, in principal, seasonal, within a specific period of time, the only way to allow for a quick return of investment (ROI)—which is the most convincing argument for a grower to invest—is by adding value to the robot by allowing it to perform more than one task, which explains why multi-purpose robots have been recently brought into the spotlight. The extent to which the ROI and labor can affect the development of robots is apparent, especially when comparing the number of commercial robots for each task and the number of robots developed by academia. Companies focus on the two operations mentioned above, weeding and harvesting, while academia focuses mostly on crop scouting, weeding, harvesting and multipurpose. Harvesting and multipurpose tasks usually require a complicated solution that attracts a lot of attention, and therefore provides exposure of the university or research center to a wider part of the population. All of the above can be clearly seen in Figure 4.
In all these operations there are specific issues that have to be tackled. Regarding the communication issue, which applies in all agricultural robots, the focus is on the way data is transmitted between the various components or the way this information reaches the user. The most popular technologies that address this challenge are WiFi [22,37,56,70,78], Ethernet [63,73,75,83,107,123], CAN-bus [13,22,27,38,75,108,153], and RS-232 [59,63,71,85,134]. For seeding robots, the optimal working speed according to crop and field features, and the image detection accuracy, are the key factors that should be further improved. A variety of sensors have been used to increase precision and speed in the seeding operation, such as RGB cameras, force, displacement and pressure, infrared sensors, and wheel encoders. However, there is still research to be carried out for better fusion of sensors, and potentially novel sensors and techniques should be tested, e.g., spectroscopy to estimate soil texture for the optimum seed depth and space placement. Weeding robots (chemical or mechanical) have been widely developed in the last years, and they offer the most commercially available systems, as weeding is tedious and labor-intensive work, especially for weeding within the crop row. However, there are optimization issues that should be faced, i.e., robots provide a small working width and slow working speed, restricting the deployment in large-scale farms and higher accuracy weed detection for selective spot spraying. So far, weeding robots have been using, in the vast majority, RGB cameras in combination with a sensor to identify the distance from the object, such as optical and acoustic distance sensors, laser range finders or stereo vision systems. Weeding robots should also be assessed and customized to be able to operate in crops with similar operational requirements.
For disease and insect detection robots, three core issues have been identified: the lack of image databases; the slow image processing (for large volumes of images); and the non-uniformity of field lighting conditions. However, these issues are gradually addressed by the use of available open-access image databases, by applying new methods on data synthesis processing, and by the use of innovative lighting systems. Similar illumination-related issues have been presented in monitoring and phenotyping robots, and they are faced by artificial lighting systems. Another challenge for the monitoring and phenotyping robots is the separation of the background. This issue has been addressed by using wider rows in the field, or by implementing other technological approaches, such as imaging of the same plant from multiple angles, and the development of better and faster algorithms. It is obvious that the amount of generated data should be efficiently handled by suitable dataset processing, automated workflows and pipelines. Spraying robots have presented similar issues to disease detection robots. Additional problems are the low spraying tank capacity and the autonomy of the robot. Even though the working speed may be considerably low, as not all plants need to be sprayed, the total spraying time of selective sprayers can be lower than conventional non-selective ones. In addition, selective sprayers can reduce drift and runoff using micro sprayers. In this light, suggested solutions include more spray nozzles, multiple operating platforms and, of course, increased speed of detection algorithms. All the above-mentioned task-specific robots use optical sensors to be able to perform their tasks, mainly RGB cameras, multi-/hyper-spectral cameras, and LiDAR sensors, therefore as optics technology advances, so will the performance of the robots.
Harvesting robots have gained wide interest globally, even though they should be further optimized. Two major problems have been underlined: the localization of the object to be harvested because leaves, fruits and other parts of the plant obstruct the view of the robot; and the potential damage to the fruit due to the sensitive nature of agricultural products. In order to develop harvesting robots with performance comparable to human workers, they should be equipped with better software and hardware; better software to deal with clusters of fruit and occlusions, and better hardware to operate harvesting reliably, without causing damages to the product and the plant. In order to harvest the ripe fruit, a vision system is needed to evaluate the maturity, and locate and grasp the fruit, and as a result, all harvesting robots include a camera (RGB, Time-of-Flight, 3D or Stereo camera) and an additional sensor to measure distance if the camera cannot provide that, such as a LiDAR sensor. Moreover, faster and more accurate vision algorithms should be developed, that work together with soft, conflict-free end effectors, as well as manipulators suitable for trees.
For the plant management robots, due to the limited number of references for each task, the challenges that have to be dealt with in the future cannot be outlined and summarized. However, given the challenges that other task-specific robots face, it could be stated that both software (e.g., better machine vision algorithms) and hardware (e.g., lightweight platforms and tools) development are needed in order to improve current robots’ performance. Moreover, further scientific research and experimental trials should be held, targeting the development of new robotic solutions in other agricultural tasks where no robotic systems exist.
Multipurpose robots perform more than one field operation, and up to now, there have been no solutions for achieving the required level of modularity and good performance. Besides those problems and solutions that have already been mentioned in the single-task robots, multipurpose robots have to perform adequately/accurately for each task they have been developed for. More tasks means more complexity hardware-wise and software-wise, and this should be a tech improvement for future robots. Because of the need to perform more than one task, the sensors used by these multipurpose robots should have a universal value; this leads to the wide use of optical sensors, such as IR and RGB cameras, in combination with more task-specific ones, like humidity, temperature and luminosity sensors. Generally, the ability to fulfill more than one task increases the cost of the robot, due to complexity of configurations and the variety of components needed. A way to justify the higher costs is by extending their functionalities and by making them compatible with various groups of crops.
Agricultural robots are associated with high investment costs from the farmer’s point of view. Consequently, there is a need for new business models in agriculture, such as robotic services contractors. As has been mentioned above, a way to return the investment fast is by investing in multi-purpose agricultural robots. In addition to this, another way to enable a quick ROI is to target high-value crops that pay back the amount invested. Furthermore, the crops that the robot will operate must require extensive crop care in order to justify replacing humans. As a result, most of the robots have been developed under this scope, with strawberries and grapes being the most popular amongst the agro-robotics community (Figure 5).
Agriculture is one of the more unstructured environments, with constant changes, and most systems are not designed in an adaptive way; therefore it is hard to robotize/automate tasks. Moreover, only recently have learning systems been able to adapt to new situations, including agriculture. Plant growth cannot be predicted accurately, obstacles may appear suddenly (animals, rocks, etc.), and occlusions are a common problem due to the growth of the plant and movement of branches. Finally, lighting conditions are constantly changing, due to clouds and the day/night cycle. As a result, one would assume that robots operating in a semi-structured environment, such as greenhouses or orchards and vineyards, would be developed as front-runners. Nevertheless, that is not the case, as almost half of the developed robots focus on open field crops (Figure 6). This could be attributed to the fact that the majority of the crops worldwide are cultivated as open field crops, or to the fact that weeding, which is one of the most popular robot categories, is mainly an issue in open fields and not greenhouses.

5. Conclusions

Overall, throughout all the reviewed robots, the most-studied robot categories were the harvesting and weeding robots, while the least-studied were disease detection and seeding robots. From the studied robots, specific and universal conclusions can be pointed out in order to address the current key issues and potential future challenges. A general overview of an agricultural robot includes part or all of the following components: a navigation system; a vision system; a control system; communication components; the robotic arm with its components; and, of course, a computer, a safety system, a remote assistance/tele-robotics system (to assist the robot when in failure mode), an edge/cloud-AI adaptive learning system and a farmer-friendly brilliant simple interface system. The majority of the robots make use of RTK-GPS for navigation, and LiDAR in order to reconstruct the environment digitally. Moreover, RGB, stereo and multi- and hyper-spectral cameras have been widely used, to allow the localization of important environmental objects and features, so that the robot can perform its task consistently and with high accuracy. Better vision systems are needed for all operations in parallel with faster image processing methods/algorithms. In particular, image processing is directly connected with processing methods, communication means, vision systems and the volume of the images. A key challenge related to the robotic vision systems is to provide uniform lighting conditions by artificial illumination means. Furthermore, new deep learning methods could be developed to improve the performance of the robotic arm end-effectors, and, more specifically, improve the eye-hand coordination of the robot by using learning methods (such as Grasp Quality Convolutional Neural Network). As for the computer hardware, it should be able to deal with the huge amount of data quickly, especially where high-resolution images are needed.

Author Contributions

S.F. contributed in the conceptualization, methodology, writing—review and editing, supervision, funding acquisition. N.M. contributed in the methodology, investigation, writing—original draft preparation, I.M. contributed in the methodology, investigation, writing—original draft preparation. E.R. contributed in the investigation, writing—original draft preparation, writing—review and editing. C.H.S. contributed to writing—review and editing, funding acquisition. E.P. contributed to writing—review and editing, funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European Commission H2020 “AgROBOfarm” project “Business-Oriented Support to the European Robotics and Agri-food Sector, towards a network of Digital Innovation Hubs in Robotics”, Grant Contract number 825395 https://agrobofood.eu/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Marinoudi, V.; Sørensen, C.G.; Pearson, S.; Bochtis, D. Robotics and labour in agriculture. A context consideration. Biosyst. Eng. 2019, 184, 111–121. [Google Scholar] [CrossRef]
  2. Pedersen, S.M.; Fountas, S.; Have, H.; Blackmore, B.S. Agricultural robots—System analysis and economic feasibility. Precis. Agric. 2006, 7, 295–308. [Google Scholar] [CrossRef]
  3. Pedersen, S.M.; Fountas, S.; Sørensen, C.G.; Van Evert, F.K.; Blackmore, B.S. Robotic seeding: Economic perspectives. In Precision Agriculture: Technology and Economic Perspectives; Pedersen, S.M., Lind, K.M., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 167–179. ISBN 978-3-319-68715-5. [Google Scholar]
  4. Blackmore, B.S.; Fountas, S.; Gemtos, T.A.; Griepentrog, H.W. A specification for an autonomous crop production mechanization system. In Proceedings of the International Symposium on Application of Precision Agriculture for Fruits and Vegetables, Orlando, FL, USA, 1 April 2009; Volume 824, pp. 201–216. [Google Scholar]
  5. Fountas, S.; Gemtos, T.A.; Blackmore, S. Robotics and Sustainability in Soil Engineering. In Soil Engineering; Springer: Berlin, Germany, 2010; pp. 69–80. [Google Scholar]
  6. Fountas, S.; Blackmore, B.; Vougioukas, S.; Tang, L.; Sørensen, C.; Jørgensen, R. Decomposition of agricultural tasks into robotic behaviours. Available online: https://cigrjournal.org/index.php/Ejounral/article/viewFile/901/895 (accessed on 6 May 2020).
  7. Reina, G.; Milella, A.; Galati, R. Terrain assessment for precision agriculture using vehicle dynamic modelling. Biosyst. Eng. 2017, 162, 124–139. [Google Scholar] [CrossRef]
  8. Fernandes, H.R.; Garcia, A.P. Design and control of an active suspension system for unmanned agricultural vehicles for field operations. Biosyst. Eng. 2018, 174, 107–114. [Google Scholar] [CrossRef]
  9. Bochtis, D.D.; Vougioukas, S.G.; Griepentrog, H.W. A mission planner for an autonomous tractor. Trans. ASABE 2009, 52, 1429–1440. [Google Scholar] [CrossRef]
  10. Bochtis, D.; Griepentrog, H.W.; Vougioukas, S.; Busato, P.; Berruto, R.; Zhou, K. Route planning for orchard operations. Comput. Electron. Agric. 2015, 113, 51–60. [Google Scholar] [CrossRef]
  11. Yang, L.; Noguchi, N. Human detection for a robot tractor using omni-directional stereo vision. Comput. Electron. Agric. 2012, 89, 116–125. [Google Scholar] [CrossRef]
  12. Noguchi, N.; Will, J.; Reid, J.; Zhang, Q. Development of a master-slave robot system for farm operations. Comput. Electron. Agric. 2004, 44, 1–19. [Google Scholar] [CrossRef]
  13. Zhang, C.; Noguchi, N.; Yang, L. Leader-follower system using two robot tractors to improve work efficiency. Comput. Electron. Agric. 2016, 121, 269–281. [Google Scholar] [CrossRef]
  14. Bechar, A.; Vigneault, C. Agricultural robots for field operations. Part 2: Operations and systems. Biosyst. Eng. 2017, 153, 110–128. [Google Scholar] [CrossRef]
  15. Aravind, K.R.; Raja, P.; Pérez-Ruiz, M. Task-based agricultural mobile robots in arable farming: A review. Span. J. Agric. Res. 2017, 15, 1–16. [Google Scholar] [CrossRef]
  16. Shamshiri, R.R.; Weltzien, C.; Hameed, I.A.; Yule, I.J.; Grift, T.E.; Balasundram, S.K.; Pitonakova, L.; Ahmad, D.; Chowdhary, G. Research and development in agricultural robotics: A perspective of digital farming. Int. J. Agric. Biol. Eng. 2018, 11, 1–14. [Google Scholar] [CrossRef]
  17. Tsolakis, N.; Bechtsis, D.; Bochtis, D. Agros: A robot operating system based emulation tool for agricultural robotics. Agronomy 2019, 9, 403. [Google Scholar] [CrossRef] [Green Version]
  18. Slaughter, D.C.; Giles, D.K.; Downey, D. Autonomous robotic weed control systems: A review. Comput. Electron. Agric. 2008, 61, 63–78. [Google Scholar] [CrossRef]
  19. Defterli, S.G.; Shi, Y.; Xu, Y.; Ehsani, R. Review of robotic technology for strawberry production. Appl. Eng. Agric. 2016, 32, 301–318. [Google Scholar]
  20. FAO Recommendations for improved weed management. Plant Prod. Prot. Div. Rome 2006, 1–56.
  21. Oerke, E.C. Crop losses to pests. J. Agric. Sci. 2006, 144, 31–43. [Google Scholar] [CrossRef]
  22. Utstumo, T.; Urdal, F.; Brevik, A.; Dørum, J.; Netland, J.; Overskeid, Ø.; Berge, T.W.; Gravdahl, J.T. Robotic in-row weed control in vegetables. Comput. Electron. Agric. 2018, 154, 36–45. [Google Scholar] [CrossRef]
  23. Dino. Available online: https://www.naio-technologies.com/en/agricultural-equipment/large-scale-vegetable-weeding-robot/ (accessed on 6 May 2020).
  24. Bakker, T.; Van Asselt, K.; Bontsema, J.; Müller, J.; Van Straten, G. An autonomous weeding robot for organic farming. In Field and Service Robotics; Springer: Berlin, Germany, 2006; Volume 25, pp. 579–590. [Google Scholar]
  25. Kim, G.H.; Kim, S.C.; Hong, Y.K.; Han, K.S.; Lee, S.G. A robot platform for unmanned weeding in a paddy field using sensor fusion. In Proceedings of the IEEE International Conference on Automation Science and Engineering, Seoul, Korea, 20–24 August 2012; pp. 904–907. [Google Scholar]
  26. Vitirover. Available online: https://www.vitirover.fr/en-home (accessed on 6 May 2020).
  27. Jørgensen, R.N.; Sørensen, C.G.; Pedersen, J.M.; Havn, I.; Jensen, K.; Søgaard, H.T.; Sorensen, L.B. Hortibot: A system design of a robotic tool carrier for high-tech plant nursing. Available online: https://ecommons.cornell.edu/bitstream/handle/1813/10601/ATOE%2007%20006%20Jorgensen%2011July2007.pdf?sequence=1&isAllowed=y (accessed on 6 May 2020).
  28. Bochtis, D.D.; Sørensen, C.G.; Jørgensen, R.N.; Nørremark, M.; Hameed, I.A.; Swain, K.C. Robotic weed monitoring. Acta Agric. Scand. Sect. B Soil Plant Sci. 2011, 61, 201–208. [Google Scholar] [CrossRef]
  29. OZ. Available online: https://www.naio-technologies.com/en/agricultural-equipment/weeding-robot-oz/ (accessed on 6 May 2020).
  30. Reiser, D.; Sehsah, E.S.; Bumann, O.; Morhard, J.; Griepentrog, H.W. Development of an autonomous electric robot implement for intra-row weeding in vineyards. Agriculture 2019, 9, 18. [Google Scholar] [CrossRef] [Green Version]
  31. Gokul, S.; Dhiksith, R.; Sundaresh, S.A.; Gopinath, M. Gesture controlled wireless agricultural weeding robot. In Proceedings of the 2019 5th International Conference on Advanced Computing and Communication Systems, San Mateo, CA, USA, 4–8 November 2019; pp. 926–929. [Google Scholar]
  32. TED. Available online: https://www.naio-technologies.com/en/agricultural-equipment/vineyard-weeding-robot/ (accessed on 6 May 2020).
  33. Anatis. Available online: https://www.carre.fr/entretien-des-cultures-et-prairies/anatis/?lang=en (accessed on 6 May 2020).
  34. Tertill. Available online: https://www.franklinrobotics.com (accessed on 6 May 2020).
  35. Asterix. Available online: https://www.adigo.no/portfolio/asterix/?lang=en (accessed on 6 May 2020).
  36. Van Evert, F.K.; Van Der Heijden, G.W.A.M.; Lotz, L.A.P.; Polder, G.; Lamaker, A.; De Jong, A.; Kuyper, M.C.; Groendijk, E.J.K.; Neeteson, J.J.; Van Der Zalm, T. A mobile field robot with vision-based detection of volunteer potato plants in a corn crop. Weed Technol. 2006, 20, 853–861. [Google Scholar] [CrossRef]
  37. EcoRobotix. Available online: https://www.ecorobotix.com/en/ (accessed on 6 May 2020).
  38. Klose, R.; Thiel, M.; Ruckelshausen, A.; Marquering, J. Weedy—A sensor fusion based autonomous field robot for selective weed control. In Proceedings of the Conference: Agricultural Engineering—Land-Technik 2008: Landtechnik regional und International, Stuttgart-Hohenheim, Germany, 25–26 September 2008; pp. 167–172. [Google Scholar]
  39. Ibex. Available online: http://www.ibexautomation.co.uk/ (accessed on 6 May 2020).
  40. BlueRiver. Available online: http://www.bluerivertechnology.com (accessed on 6 May 2020).
  41. RIPPA & VIPPA. Available online: https://confluence.acfr.usyd.edu.au/display/AGPub/Our+Robots (accessed on 6 May 2020).
  42. Lee, W.S.; Slaughter, D.C.; Giles, D.K. Robotic weed control system for tomatoes. Precis. Agric. 1999, 1, 95–113. [Google Scholar] [CrossRef]
  43. Testa, G.; Reyneri, A.; Blandino, M. Maize grain yield enhancement through high plant density cultivation with different inter-row and intra-row spacings. Eur. J. Agron. 2016, 72, 28–37. [Google Scholar] [CrossRef]
  44. Haibo, L.; Shuliang, D.; Zunmin, L.; Chuijie, Y. Study and experiment on a wheat precision seeding robot. J. Robot. 2015. [Google Scholar] [CrossRef]
  45. Ruangurai, P.; Ekpanyapong, M.; Pruetong, C.; Watewai, T. Automated three-wheel rice seeding robot operating in dry paddy fields. Maejo Int. J. Sci. Technol. 2015, 9, 403–412. [Google Scholar]
  46. Sunitha, K.A.; Suraj, G.S.G.S.; Sowrya, C.H.P.N.; Sriram, G.A.; Shreyas, D.; Srinivas, T. Agricultural robot designed for seeding mechanism. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2017; Volume 197. [Google Scholar]
  47. Xaver. Available online: https://www.fendt.com/int/xaver (accessed on 6 May 2020).
  48. Mars. Available online: https://www.fendt.com/int/fendt-mars (accessed on 6 May 2020).
  49. Katupitiya, J. An autonomous seeder for broad acre crops. In Proceedings of the American Society of Agricultural and Biological Engineers Annual International Meeting 2014, St Joseph, MI, USA, 13–16 July 2014; Volume 1, pp. 169–176. [Google Scholar]
  50. Rehman, T.U.; Mahmud, M.S.; Chang, Y.K.; Jin, J.; Shin, J. Current and future applications of statistical machine learning algorithms for agricultural machine vision systems. Comput. Electron. Agric. 2019, 156, 585–605. [Google Scholar] [CrossRef]
  51. Schor, N.; Bechar, A.; Ignat, T.; Dombrovsky, A.; Elad, Y.; Berman, S. Robotic disease detection in greenhouses: Combined detection of powdery mildew and tomato spotted wilt virus. IEEE Robot. Autom. Lett. 2016, 1, 354–360. [Google Scholar] [CrossRef]
  52. Schor, N.; Berman, S.; Dombrovsky, A.; Elad, Y.; Ignat, T.; Bechar, A. Development of a robotic detection system for greenhouse pepper plant diseases. Precis. Agric. 2017, 18, 394–409. [Google Scholar] [CrossRef]
  53. Rey, B.; Aleixos, N.; Cubero, S.; Blasco, J. XF-ROVIM. A field robot to detect olive trees infected by Xylella fastidiosa using proximal sensing. Remote Sens. 2019, 11, 221. [Google Scholar] [CrossRef] [Green Version]
  54. Sultan Mahmud, M.; Zaman, Q.U.; Esau, T.J.; Price, G.W.; Prithiviraj, B. Development of an artificial cloud lighting condition system using machine vision for strawberry powdery mildew disease detection. Comput. Electron. Agric. 2019, 158, 219–225. [Google Scholar] [CrossRef]
  55. Liu, B.; Hu, Z.; Zhao, Y.; Bai, Y.; Wang, Y. Recognition of pyralidae insects using intelligent monitoring autonomous robot vehicle in natural farm scene. arXiv 2019, arXiv:physics/1903.10827. [Google Scholar]
  56. Pilli, S.K.; Nallathambi, B.; George, S.J.; Diwanji, V. EAGROBOT—A robot for early crop disease detection using image processing. In Proceedings of the 2nd International Conference on Electronics and Communication Systems, Coimbatore, India, 26–27 February 2015; pp. 1684–1689. [Google Scholar]
  57. Zheng, Y.-Y.; Kong, J.-L.; Jin, X.-B.; Wang, X.-Y.; Su, T.-L.; Zuo, M. CropDeep: The crop vision dataset for deep-learning-based classification and detection in precision agriculture. Sensors 2019, 19, 1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Barth, R.; IJsselmuiden, J.; Hemming, J.; Van Henten, E.J. Data synthesis methods for semantic segmentation in agriculture: A Capsicum annuum dataset. Comput. Electron. Agric. 2018, 144, 284–296. [Google Scholar] [CrossRef]
  59. Bayati, M.; Fotouhi, R. A mobile robotic platform for crop monitoring. Adv. Robot. Autom. 2018, 7, 2. [Google Scholar] [CrossRef]
  60. Xue, J.; Su, B. Significant remote sensing vegetation indices: A review of developments and applications. Sensors 2017. [Google Scholar] [CrossRef] [Green Version]
  61. Wang, Y.; Hu, X.; Hou, Z.; Ning, J.; Zhang, Z. Discrimination of nitrogen fertilizer levels of tea plant (Camellia sinensis) based on hyperspectral imaging. J. Sci. Food Agric. 2018, 98, 4659–4664. [Google Scholar] [CrossRef]
  62. Bietresato, M.; Carabin, G.; Vidoni, R.; Gasparetto, A.; Mazzetto, F. Evaluation of a LiDAR-based 3D-stereoscopic vision system for crop-monitoring applications. Comput. Electron. Agric. 2016, 124, 1–13. [Google Scholar] [CrossRef]
  63. Vidoni, R.; Gallo, R.; Ristorto, G.; Carabin, G.; Mazzetto, F.; Scalera, L.; Gasparetto, A. Byelab: An agricultural mobile robot prototype for proximal sensing and precision farming. In Proceedings of the ASME International Mechanical Engineering Congress and Exposition, Tampa, FL, USA, 3–9 November 2017; Volume 4A-2017. [Google Scholar]
  64. OUVA. Available online: https://www.polariks.com/text?rq=ouva (accessed on 6 May 2020).
  65. Vinbot. Available online: http://vinbot.eu/ (accessed on 6 May 2020).
  66. Vinescout. Available online: http://vinescout.eu/web/ (accessed on 6 May 2020).
  67. Grape. Available online: http://www.grape-project.eu/home/ (accessed on 6 May 2020).
  68. Dos Santos, F.N.; Sobreira, H.; Campos, D.; Morais, R.; Moreira, A.P.; Contente, O. Towards a reliable monitoring robot for mountain vineyards. In Proceedings of the 2015 IEEE International Conference on Autonomous Robot Systems and Competitions, ICARSC 2015, Vila Real, Portugal, 8–10 April 2015; pp. 37–43. [Google Scholar]
  69. Terrasentia. Available online: https://www.earthsense.co (accessed on 6 May 2020).
  70. Vaeljaots, E.; Lehiste, H.; Kiik, M.; Leemet, T. Soil sampling automation case-study using unmanned ground vehicle. Eng. Rural Dev. 2018, 17, 982–987. [Google Scholar]
  71. Iida, M.; Kang, D.; Taniwaki, M.; Tanaka, M.; Umeda, M. Localization of CO2 source by a hexapod robot equipped with an anemoscope and a gas sensor. Comput. Electron. Agric. 2008, 63, 73–80. [Google Scholar] [CrossRef]
  72. Deery, D.; Jimenez-Berni, J.; Jones, H.; Sirault, X.; Furbank, R. Proximal remote sensing buggies and potential applications for field-based phenotyping. Agronomy 2014, 4, 349–379. [Google Scholar] [CrossRef] [Green Version]
  73. Mueller-Sim, T.; Jenkins, M.; Abel, J.; Kantor, G. The Robotanist: A ground-based agricultural robot for high-throughput crop phenotyping. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 July 2017; pp. 3634–3639. [Google Scholar]
  74. Benet, B.; Dubos, C.; Maupas, F.; Malatesta, G.; Lenain, R. Development of autonomous robotic platforms for sugar beet crop phenotyping using artificial vision. Available online: https://hal.archives-ouvertes.fr/hal-02155159/document (accessed on 6 May 2020).
  75. Ruckelshausen, A.; Biber, P.; Dorna, M.; Gremmes, H.; Klose, R.; Linz, A.; Rahe, R.; Resch, R.; Thiel, M.; Trautz, D.; et al. BoniRob: An autonomous field robot platform for individual plant phenotyping. In Proceedings of the Precision Agriculture 2009—Papers Presented at the 7th European Conference on Precision Agriculture, Wageningen, The Netherlands, 6–8 July 2009; pp. 841–847. [Google Scholar]
  76. Virlet, N.; Sabermanesh, K.; Sadeghi-Tehran, P.; Hawkesford, M.J. Field Scanalyzer: An automated robotic field phenotyping platform for detailed crop monitoring. Funct. Plant Biol. 2017, 44, 143–153. [Google Scholar] [CrossRef] [Green Version]
  77. Bao, Y.; Tang, L.; Breitzman, M.W.; Salas Fernandez, M.G.; Schnable, P.S. Field-based robotic phenotyping of sorghum plant architecture using stereo vision. J. Field Robot. 2019, 36, 397–415. [Google Scholar] [CrossRef]
  78. Shafiekhani, A.; Kadam, S.; Fritschi, F.B.; Desouza, G.N. Vinobot and vinoculer: Two robotic platforms for high-throughput field phenotyping. Sensors 2017, 17, 214. [Google Scholar] [CrossRef] [PubMed]
  79. Young, S.N.; Kayacan, E.; Peschel, J.M. Design and field evaluation of a ground robot for high-throughput phenotyping of energy sorghum. Precis. Agric. 2019, 20, 697–722. [Google Scholar] [CrossRef] [Green Version]
  80. Zhang, S.; Wang, H.; Huang, W.; You, Z. Plant diseased leaf segmentation and recognition by fusion of superpixel, K-means and PHOG. Optik 2018, 157, 866–872. [Google Scholar] [CrossRef]
  81. Vacavant, A.; Lebre, M.A.; Rositi, H.; Grand-Brochier, M.; Strand, R. New definition of quality-scale robustness for image processing algorithms, with generalized uncertainty modeling, applied to denoising and segmentation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Beijing, China, 2019; Volume 11455 LNCS, pp. 138–149. [Google Scholar]
  82. Rajbhoj, A.; Deshpande, S.; Gubbi, J.; Kulkarni, V.; Balamuralidhar, P. A System for semi-automatic construction of image processing pipeline for complex problems. In fs have already been under development over the last years, and there are also a number of commercial robots on the market. Most of the major problems related tg; Springer: Rome, Italy, 2019; Volume 352, pp. 295–310. [Google Scholar]
  83. Sammons, P.J.; Furukawa, T.; Bulgin, A. Autonomous pesticide spraying robot for use in a greenhouse. In Proceedings of the 2005 Australasian Conference on Robotics and Automation, Sydney, Australia, 5–7 December 2005. [Google Scholar]
  84. Holland Green Machine. Available online: http://www.hollandgreenmachine.com/sprayrobot/product-information-spray-robot-s55/ (accessed on 6 May 2020).
  85. Sanchez-Hermosilla, J.; Rodriguez, F.; Gonzalez, R.; Luis, J.; Berenguel, M. A mechatronic description of an autonomous mobile robot for agricultural tasks in greenhouses. In Mobile Robots Navigation; INTECH Open Access Publisher: London, UK, 2010. [Google Scholar]
  86. Mahmud, M.S.A.; Abidin, M.S.Z.; Mohamed, Z.; Rahman, M.K.I.A.; Iida, M. Multi-objective path planner for an agricultural mobile robot in a virtual greenhouse environment. Comput. Electron. Agric. 2019, 157, 488–499. [Google Scholar] [CrossRef]
  87. Singh, S.; Burks, T.F.; Lee, W.S. Autonomous robotic vehicle development for greenhouse spraying. Trans. Am. Soc. Agric. Eng. 2005, 48, 2355–2361. [Google Scholar] [CrossRef]
  88. Oberti, R.; Marchi, M.; Tirelli, P.; Calcante, A.; Iriti, M.; Hočevar, M.; Baur, J.; Pfaff, J.; Schütz, C.; Ulbrich, H. Selective spraying of grapevine’s diseases by a modular agricultural robot. J. Agric. Eng. 2013, 44, 149–153. [Google Scholar] [CrossRef]
  89. Underwood, J.P.; Calleija, M.; Taylor, Z.; Hung, C.; Nieto, J.; Fitch, R.; Sukkarieh, S. Real-time target detection and steerable spray for vegetable crops. In Proceedings of the International Conference on Robotics and Automation: Robotics in Agriculture Workshop, Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
  90. Søgaard, H.T.; Lund, I. Application accuracy of a machine vision-controlled robotic micro-dosing system. Biosyst. Eng. 2007, 96, 315–322. [Google Scholar] [CrossRef]
  91. Ogawa, Y.; Kondo, N.; Monta, M.; Shibusawa, S. Spraying robot for grape production. Springer Tracts Adv. Robot. 2006, 24, 539–548. [Google Scholar]
  92. Hayashi, S.; Yamamoto, S.; Tsubota, S.; Ochiai, Y.; Kobayashi, K.; Kamata, J.; Kurita, M.; Inazumi, H.; Peter, R. Automation technologies for strawberry harvesting and packing operations in Japan. J. Berry Res. 2014, 4, 19–27. [Google Scholar] [CrossRef] [Green Version]
  93. Feng, Q.; Zheng, W.; Qiu, Q.; Jiang, K.; Guo, R. Study on strawberry robotic harvesting system. In Proceedings of the CSAE 2012 IEEE International Conference on Computer Science and Automation Engineering, Zhangjiajie, China, 25–27 May 2012; Volume 1, pp. 320–324. [Google Scholar]
  94. Qingchun, F.; Xiu, W.; Wengang, Z.; Quan, Q.; Kai, J. New strawberry harvesting robot for elevated-trough culture. Int. J. Agric. Biol. Eng. 2012, 5, 1–8. [Google Scholar]
  95. Xiong, Y.; Peng, C.; Grimstad, L.; From, P.J.; Isler, V. Development and field evaluation of a strawberry harvesting robot with a cable-driven gripper. Comput. Electron. Agric. 2019, 157, 392–402. [Google Scholar] [CrossRef]
  96. Harvest Croo Robotics. Available online: https://harvestcroo.com/ (accessed on 6 May 2020).
  97. Arima, S.; Kondo, N.; Monta, M. Strawberry harvesting robot on table-top culture. In 2004 ASAE Annual Meeting; American Society of Agricultural and Biological Engineers: San Jose, MI, USA, 2004; p. 1. [Google Scholar]
  98. Hayashi, S.; Shigematsu, K.; Yamamoto, S.; Kobayashi, K.; Kohno, Y.; Kamata, J.; Kurita, M. Evaluation of a strawberry-harvesting robot in a field test. Biosyst. Eng. 2010, 105, 160–171. [Google Scholar] [CrossRef]
  99. Shiigi, T.; Kondo, N.; Kurita, M.; Ninomiya, K.; Rajendra, P.; Kamata, J.; Hayashi, S.; Kobayashi, K.; Shigematsu, K.; Kohno, Y. Strawberry harvesting robot for fruits grown on table top culture. In Proceedings of the American Society of Agricultural and Biological Engineers Annual International Meeting 2008, Providence, RI, USA, 29 June–2 July 2008; Volume 5, pp. 3139–3148. [Google Scholar]
  100. Dogtooth. Available online: https://dogtooth.tech/ (accessed on 6 May 2020).
  101. Agrobot E-Series. Available online: http://agrobot.com/ (accessed on 6 May 2020).
  102. Octinion. Available online: http://octinion.com/products/agricultural-robotics/rubion (accessed on 6 May 2020).
  103. Silwal, A.; Davidson, J.; Karkee, M.; Mo, C.; Zhang, Q.; Lewis, K. Effort towards robotic apple harvesting in Washington State. In Proceedings of the 2016 American Society of Agricultural and Biological Engineers Annual International Meeting, Orlando, FL, USA, 17–20 July 2016. [Google Scholar]
  104. Baeten, J.; Donné, K.; Boedrij, S.; Beckers, W.; Claesen, E. Autonomous fruit picking machine: A robotic apple harvester. In Field and service robotics; Springer: Berlin, Germany, 2008; Volume 42, pp. 531–539. [Google Scholar]
  105. Bulanon, D.; Kataoka, T. Fruit detection system and an end effector for robotic harvesting of Fuji apples. Agric. Eng. Int. CIGR J. 2010, 12, 203–210. [Google Scholar]
  106. FF Robotics. Available online: https://www.ffrobotics.com/ (accessed on 6 May 2020).
  107. Yaguchi, H.; Nagahama, K.; Hasegawa, T.; Inaba, M. Development of an autonomous tomato harvesting robot with rotational plucking gripper. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Daejeon, Korea, 9–14 October 2016; pp. 652–657. [Google Scholar]
  108. Wang, L.L.; Zhao, B.; Fan, J.W.; Hu, X.A.; Wei, S.; Li, Y.S.; Zhou, Q.B.; Wei, C.F. Development of a tomato harvesting robot used in greenhouse. Int. J. Agric. Biol. Eng. 2017, 10, 140–149. [Google Scholar]
  109. Feng, Q.; Zou, W.; Fan, P.; Zhang, C.; Wang, X. Design and test of robotic harvesting system for cherry tomato. Int. J. Agric. Biol. Eng. 2018, 11, 96–100. [Google Scholar] [CrossRef]
  110. Metomotion. Available online: https://metomotion.com (accessed on 6 May 2020).
  111. Root-AI. Available online: https://root-ai.com (accessed on 6 May 2020).
  112. Energid. Available online: https://www.energid.com/industries/agricultural-robotics (accessed on 6 May 2020).
  113. Tanigaki, K.; Fujiura, T.; Akase, A.; Imagawa, J. Cherry-harvesting robot. Comput. Electron. Agric. 2008, 63, 65–72. [Google Scholar] [CrossRef]
  114. Ceres, R.; Pons, J.L.; Jiménez, A.R.; Martín, J.M.; Calderón, L. Design and implementation of an aided fruit-harvesting robot (Agribot). Ind. Rob. 1998, 25, 337–346. [Google Scholar] [CrossRef]
  115. Van Henten, E.J.; Hemming, J.; Van Tuijl, B.A.J.; Kornet, J.G.; Meuleman, J.; Bontsema, J.; Van Os, E.A. An autonomous robot for harvesting cucumbers in greenhouses. Auton. Robot. 2002, 13, 241–258. [Google Scholar] [CrossRef]
  116. Hayashi, S.; Ganno, K.; Ishii, Y.; Tanaka, I. Robotic harvesting system for eggplants. Jpn. Agric. Res. Q. 2002, 36, 163–168. [Google Scholar] [CrossRef] [Green Version]
  117. Sweeper. Available online: http://www.sweeper-robot.eu (accessed on 6 May 2020).
  118. Irie, N.; Taguchi, N.; Horie, T.; Ishimatsu, T. Asparagus harvesting robot coordinated with 3-D vision sensor. In Proceedings of the IEEE International Conference on Industrial Technology, Gippsland, Australia, 10–13 February 2009. [Google Scholar]
  119. Cerescon. Available online: https://www.cerescon.com/EN/sparter (accessed on 6 May 2020).
  120. Roshanianfard, A.R.; Noguchi, N. Kinematics analysis and simulation of A 5DOF articulated robotic arm applied to heavy products harvesting. Tarım Bilim. Derg. 2018, 24, 91–104. [Google Scholar] [CrossRef] [Green Version]
  121. Edan, Y.; Rogozin, D.; Flash, T.; Miles, G.E. Robotic melon harvesting. IEEE Trans. Robot. Autom. 2000, 16, 831–835. [Google Scholar] [CrossRef]
  122. Umeda, M.; Kubota, S.; Iida, M. Development of “STORK”, a watermelon-harvesting robot. Artif. Life Robot. 1999, 3, 143–147. [Google Scholar] [CrossRef]
  123. Pilarski, T.; Happold, M.; Pangels, H.; Ollis, M.; Fitzpatrick, K.; Stentz, A. The Demeter system for automated harvesting. Auton. Robots 2002, 13, 9–20. [Google Scholar] [CrossRef]
  124. Rowley, J.H. Developing Flexible Automation for Mushroom Harvesting (Agaricus Bisporus): Innovation Report; University of Warwick: Coventry, UK, 2009. [Google Scholar]
  125. Siciliano, B.; Khatib, O. Springer Handbook of Robotics; Springer: Berlin, Germany, 2016; ISBN 3319325523. [Google Scholar]
  126. Abundant Robotics. Available online: https://www.abundantrobotics.com/ (accessed on 6 May 2020).
  127. Zapotezny-Anderson, P.; Lehnert, C. Towards active robotic vision in agriculture: A deep learning approach to visual servoing in occluded and unstructured protected cropping environments. arXiv 2019, arXiv:Prepr/1908.01885. [Google Scholar] [CrossRef]
  128. Mu, L.; Cui, G.; Liu, Y.; Cui, Y.; Fu, L.; Gejima, Y. Design and simulation of an integrated end-effector for picking kiwifruit by robot. Inf. Process. Agric. 2019, 7, 58–71. [Google Scholar] [CrossRef]
  129. Harvest Automation. Available online: https://www.public.harvestai.com/ (accessed on 6 May 2020).
  130. Vision Robotics. Available online: https://www.visionrobotics.com/vr-grapevine-pruner (accessed on 6 May 2020).
  131. Botterill, T.; Paulin, S.; Green, R.; Williams, S.; Lin, J.; Saxton, V.; Mills, S.; Chen, X.; Corbett-Davies, S. A robot system for pruning grape vines. J. F. Robot. 2017, 34, 1100–1122. [Google Scholar] [CrossRef]
  132. Zahid, A.; He, L.; Zeng, L. Development of a robotic end effector for apple tree pruning. In Proceedings of the 2019 ASABE Annual International Meeting, Boston, MA, USA, 7–10 July 2019. [Google Scholar]
  133. He, L.; Zhou, J.; Zhang, Q.; Charvet, H.J. A string twining robot for high trellis hop production. Comput. Electron. Agric. 2016, 121, 207–214. [Google Scholar] [CrossRef] [Green Version]
  134. Grimstad, L.; From, P.J. The Thorvald II agricultural robotic system. Robotics 2017, 6, 24. [Google Scholar] [CrossRef] [Green Version]
  135. Xue, J.; Fan, B.; Zhang, X.; Feng, Y. An agricultural robot for multipurpose operations in a greenhouse. DEStech Trans. Eng. Technol. Res. 2017. [Google Scholar] [CrossRef] [Green Version]
  136. Narendran, V. Autonomous robot for E-farming based on fuzzy logic reasoning. Int. J. Pure Appl. Math. 2018, 118, 3811–3821. [Google Scholar]
  137. Robotti. Available online: http://agrointelli.com/robotti-diesel.html#rob.diesel (accessed on 6 May 2020).
  138. Durmus, H.; Gunes, E.O.; Kirci, M.; Ustundag, B.B. The design of general purpose autonomous agricultural mobile-robot: “AGROBOT”. In Proceedings of the 2015 4th International Conference on Agro-Geoinformatics, Agro-Geoinformatics 2015, Istanbul, Turkey, 20–24 July 2015; pp. 49–53. [Google Scholar]
  139. Peña, C.; Riaño, C.; Moreno, G. RobotGreen: A teleoperated agricultural robot for structured environments. J. Eng. Sci. Technol. Rev. 2018, 11, 87–98. [Google Scholar] [CrossRef]
  140. Murakami, N.; Ito, A.; Will, J.D.; Steffen, M.; Inoue, K.; Kita, K.; Miyaura, S. Development of a teleoperation system for agricultural vehicles. Comput. Electron. Agric. 2008, 63, 81–88. [Google Scholar] [CrossRef]
  141. Jasiński, M.; Mączak, J.; Radkowski, S.; Korczak, S.; Rogacki, R.; Mac, J.; Szczepaniak, J. Autonomous agricultural robot—Conception of inertial navigation system. In International Conference on Automation; Springer: Cham, Switzerland, 2016; Volume 440, pp. 669–679. [Google Scholar]
  142. Chang, C.L.; Lin, K.M. Smart agricultural machine with a computer vision-based weeding and variable-rate irrigation scheme. Robotics 2018, 7, 38. [Google Scholar] [CrossRef] [Green Version]
  143. Moshou, D.; Bravo, C.; Oberti, R.; West, J.S.; Ramon, H.; Vougioukas, S.; Bochtis, D. Intelligent multi-sensor system for the detection and treatment of fungal diseases in arable crops. Biosyst. Eng. 2011, 108, 311–321. [Google Scholar] [CrossRef]
  144. Quan, L.; Feng, H.; Lv, Y.; Wang, Q.; Zhang, C.; Liu, J.; Yuan, Z. Maize seedling detection under different growth stages and complex field environments based on an improved Faster R–CNN. Biosyst. Eng. 2019, 184, 1–23. [Google Scholar] [CrossRef]
  145. Gollakota, A.; Srinivas, M.B. Agribot—A multipurpose agricultural robot. In Proceedings of the 2011 Annual IEEE India Conference, Hyderabad, India, 16–18 December 2011; pp. 1–4. [Google Scholar]
  146. Praveena, R.; Srimeena, R. Agricultural robot for automatic ploughing and seeding. In Proceedings of the 2015 IEEE Technological Innovation in ICT for Agriculture and Rural Development (TIAR), Chennai, India, 10–12 July 2015; pp. 17–23. [Google Scholar]
  147. Sowjanya, K.D.; Sindhu, R.; Parijatham, M.; Srikanth, K.; Bhargav, P. Multipurpose autonomous agricultural robot. In Proceedings of the International Conference on Electronics, Communication and Aerospace Technology, Coimbatore, India, 20–22 April 2017; pp. 696–699. [Google Scholar]
  148. Ranjitha, B.; Nikhitha, M.N.; Aruna, K.; Murthy, B.T.V. Solar powered autonomous multipurpose agricultural robot using bluetooth/android app. In Proceedings of the 2019 3rd International conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 12–14 June 2019; pp. 872–877. [Google Scholar]
  149. Wall-ye (MYCE_Agriculture). Available online: http://wall-ye.com/index-6.html (accessed on 6 May 2020).
  150. Wall-ye (Myce_Vigne). Available online: http://wall-ye.com/index.html (accessed on 6 May 2020).
  151. Green, O.; Schmidt, T.; Pietrzkowski, R.P.; Jensen, K.; Larsen, M.; Jørgensen, R.N. Commercial autonomous agricultural platform. In Proceedings of the Second International Conference on Robotics and associated High-technologies and Equipment for Agriculture and Forestry, Bergamo, Italy, 20–23 October 2014; pp. 351–356. [Google Scholar]
  152. TrimBot 2020. Available online: http://trimbot2020.webhosting.rug.nl/ (accessed on 6 May 2020).
  153. Bak, T.; Jakobsen, H. Agricultural robotic platform with four wheel steering for weed detection. Biosyst. Eng. 2004, 87, 125–136. [Google Scholar] [CrossRef]
Figure 1. Steps of review methodology.
Figure 1. Steps of review methodology.
Sensors 20 02672 g001
Figure 2. Global illustration of the reviewed robots’ overall performance.
Figure 2. Global illustration of the reviewed robots’ overall performance.
Sensors 20 02672 g002
Figure 3. Representation of various types of agricultural robots ((1) weeding [23], (2) seeding [47], (3) disease and insect detection [55], (4) plant monitoring [69], (5) phenotyping [78], (6) spraying [86], (7) harvesting [100], (8) plant management [152], (9) multi-purpose [137]).
Figure 3. Representation of various types of agricultural robots ((1) weeding [23], (2) seeding [47], (3) disease and insect detection [55], (4) plant monitoring [69], (5) phenotyping [78], (6) spraying [86], (7) harvesting [100], (8) plant management [152], (9) multi-purpose [137]).
Sensors 20 02672 g003
Figure 4. Number of reviewed robots per field operation.
Figure 4. Number of reviewed robots per field operation.
Sensors 20 02672 g004
Figure 5. The main crops in correlation with the number of robotic systems.
Figure 5. The main crops in correlation with the number of robotic systems.
Sensors 20 02672 g005
Figure 6. Allocation (%) of robotic systems in various agricultural environments.
Figure 6. Allocation (%) of robotic systems in various agricultural environments.
Sensors 20 02672 g006
Table 1. Weeding agrirobotic systems’ characteristics for various crops.
Table 1. Weeding agrirobotic systems’ characteristics for various crops.
CropPerception SensorsWeed DetectionWeed ControlResultsCited Work
MaizeCameras, optical and acoustic distance sensorsYesChemicalNo performance metrics provided[38]
CarrotRGB infrared cameraPartlyChemical100% effectiveness with the DoD system[22]
Potato, cornWebcam, solid-state gyroscopePartlyChemical98% and 89% detection accuracy[36]
Sugar beetColor cameraYesMechanicalRow detection precision < 25 mm.
> 90% in-row weed removal
[24]
N/AStereo vision system, laserNoMechanicalPrecision < 3 cm.[27,29]
RiceLaser range finder, IMU NoMechanicalPrecision < 62 mm[25]
BeetrootColor camera, artificial vision, compassYesChemical> 85% detection & destroy, precision < 2 cm[37]
GrapesIMU, hall sensors, electromechanical sensor, a sonar sensorNoMechanicalAverage performance: 65% (feeler) & 82% (sonar)[30]
N/AAccelerometer, gyroscope, flex sensorNoMechanicalNo performance metrics provided[31]
TomatoColor camera, SensorWatchPartlyChemical24.2% were incorrectly identified and sprayed and 52.4% of the weeds were not sprayed.[42]
Table 2. Seeding agrirobotic systems’ characteristics for wheat and rice.
Table 2. Seeding agrirobotic systems’ characteristics for wheat and rice.
CropPerceptionResultsCited Work
WheatForce sensor, displacement sensor, angle sensorThe path tracking errors are +/− 5 cm and the angle errors are about zero. [49]
WheatSignal sensor, angle, pressure & infrared sensorsQualified rate: 93.3%[44]
RiceCompass, wheel encoder92% accuracy & 5 cm error in the dropping position. [45]
Table 3. Disease and insect detection agrirobotic systems features for various crops.
Table 3. Disease and insect detection agrirobotic systems features for various crops.
CropPerceptionDetected DiseaseHighest AccuracyCited Work
Bell pepperRGB camera, multispectral camera, laser sensorPowdery mildew & tomato spotted wilt virus 95% & 90%[51,52]
Cotton, groundnutRGB cameraCotton (Bacterial blight, magnesium deficiency), groundnut (leaf spot & anthracnose)~90%, 83–96%[56]
Olive treeTwo DSLR cameras (one in BNDVI mode), a multispectral camera, a hyperspectral system in visible and NIR range, a thermal camera, LiDAR, an IMU sensor(*) Xylella fastidiosa bacteriumN/A[53]
Tomato, riceRGB cameraPyralidae insect94.3%[55]
StrawberryRGB cameraPowdery mildew72%-95%[54]
(*)DSLR: Digital Single-lens Reflex. BNDVI: Blue Normalised Difference Vegetation Index. NIR: near-infrared. LiDAR: Light detection and ranging.
Table 4. Plant vigor monitoring agrirobotic systems for various crops.
Table 4. Plant vigor monitoring agrirobotic systems for various crops.
CropPerceptionResultsCited Work
No specific cropHigh-resolution stereo-cameras, 3D LiDARSoil sampling. No performance metrics provided.[70]
No specific cropCO2 gas sensor, anemoscope, IR distance measuring sensorGas source tracking. CO2 concentration levels up to 2500 ppm were recorded while the robot was moving at a speed of 2 m/min. [71]
OrchardsLiDAR, luxmeterCanopy volume estimation. The system is independent of the light conditions, it is highly reliable and data processing is very fast.[62]
GrapesRGB & IR camera, laser ranger finder, IMU, pressure sensor, etc.Crop monitoring tasks. No performance metrics provided.[68]
Orchards and vineyardsLiDAR, OptRX sensorMonitor health status and canopy thickness. Terrestrial Laser Scanning (TLS): 2 mm distance accuracy[63]
CanolaUltrasonic sensors, NDVI sensors, IR thermometers, RGB cameraGather phenotypic data. Maximum measurement error: 2.5%[59]
Table 5. Phenotyping agrirobotic systems for various crops.
Table 5. Phenotyping agrirobotic systems for various crops.
CropPerceptionAutonomy LevelResultsCited Work
Maize & wheatCameras, spectral imaging systems, laser sensors, 3D time-of-flight cameras AutonomousNo performance metrics provided [75]
CottonStereo RGB & IR thermal camera, temperature, humidity, light intensity sensors, pyranometer, quantum, LiDARSemi-autonomousRMS error:
Plant height: < 0.5 cm (Vinobot)
RGB to IR calibration: 2.5 px (Vinoculer)
Temp: < 1 °C (Vinoculer)
[78]
SorghumStereo camera, RGB camera with fish eye lenses, penetrometerAutonomousStalk detection: 96%[73]
Rice, maize & wheatRGB camera, chlorophyll fluorescence camera, NDVI camera, thermal infrared camera, hyperspectral camera, 3D laser scannerFixed site fully automatedPlant height RMS error: 1.88 cm[76]
Sugar BeetMobile robot: Webcam camera, gigaethernet camera
Bettybot: Color camera, hyperspectral camera
AutonomousNo performance metrics provided[74]
SorghumStereo imaging system consisting of color camerasAutonomous based on commercial tractorThe image-derived measurements were highly repeatable &
showed high correlations with manual measurements.
[77]
Energy SorghumStereo camera, time of flight depth sensor, IR cameraSemi-autonomousAverage absolute error for stem width and plant height: 13% and 15%.[79]
Table 6. Spraying agrirobotic systems’ features for various crops.
Table 6. Spraying agrirobotic systems’ features for various crops.
CropPerceptionReal-time DetectionResultsCited Work
CantaloupeRobot controllerNoNSGA-II execution time was better 1.5–7% than NSGA-III for the same test cases. (*)[86]
N/AWeb cameraYes27% off-target shots, 99.8% of the targets were sprayed by at least one shot.[90]
CucumberBump sensors, infra-red sensors, induction sensorsNoRun success:90% & 95%, topside leaf coverage:95% & 90%, underside leaf coverage:95 & 80%, over-spray:20% & 10% for Tests 1 & 2[83]
N/AUltrasonic sensorsNoMaximum error varied from 1.2–4.5 cm for self-contained mode on concrete, while for trailer varied from 2.2–4.9 cm.[87]
GrapevineUltrasonic sensor, color TV cameraNoNo performance metrics.[91]
N/AMiddle-range sonar, short-range sonar, radar, compass, radarNoLongitudinal, lateral and orientation errors are close to zero, by including slip varied 10–30%.[85]
GrapevineRGB camera, R-G-NIR multispectral cameraYesThe sensitivity of the robotic selective spraying was 85%, the selectivity was 92% (8% of the total healthy area was sprayed unnecessarily).[88]
Vegetable cropsHyperspectral camera, stereo vision, thermal IR camera, monocular color cameraYesThe greedy sort and raster methods are substantially faster than back-to-front scheduling, taking only 68% to 77% of the time.[89]
(*)NSGA: Non-dominated Sorting Genetic Algorithm.
Table 7. Harvesting agrirobotic systems’ features for various crops.
Table 7. Harvesting agrirobotic systems’ features for various crops.
CropPerceptionFastest Picking SpeedHighest Picking RateCited Work
alfalfa, sudanColor camera, gyroscope2 ha/h (alfalfa)N/A[123]
apple treeColor camera, time of flight based three dimensional camera7.5 sec/fruit84% [103]
apple treeColor CCD (Charge Coupled Device) camera, laser range sensor7.1 sec/fruit89% [105]
apple treeHigh-frequency light, camera9 sec/fruit80%[104]
cherry3D vision sensor with red, IR laser diodes, pressure sensor14 sec/fruitN/A[113]
mushroomLaser sensor, vision sensor6.7 sec/mushroom69% [124]
asparagus3D vision sensor with two sets of slit laser projectors & a TV camera13.7 sec/asparagusN/A[118]
strawberrySonar camera sensor, binocular camera31.3 sec/fruit86% [94]
strawberryColor CCD cameras, reflection-type photoelectric sensor8.6 sec/fruit54.9% [92]
strawberryLED light source, three-color CCD cameras, photoelectric sensor, suction device11.5 sec/fruit41.3% with a suction device
34.9% without it
[98]
strawberryColor CCD camera, visual sensor10 sec/fruitN/A[97]
strawberryThree VGA (Video Graphics Array) class CCD color cameras (stereo vision system and center camera)N/A46% [99]
strawberryRGB-D camera, 3 IR sensors10.6 sec/fruit53.6%[95]
tomatoStereo camera, playstation camera23 sec/tomato60% [107]
tomatoBinocular stereo vision system, laser sensor15 sec/tomato86% [108]
cherry tomatoCamera, laser sensor8 sec/tomato bunch83%[109]
cucumberTwo synchronized CCD cameras45 sec/cucumber80%[115]
various fruitsPressure sensor, 2 convergent IR sensors, telemeter, cameras2 sec/fruit (only grasp & detach)N/A[114]
melonTwo black and white CCD cameras, proximity sensor, far and near vision sensors15 sec/fruit85.67% [121]
eggplantSingle CCD camera, photoelectric sensor64.1 sec/eggplant62.5%[116]
watermelonTwo CCD cameras, vacuum sensorN/A66.7% [122]
Table 8. Plant management agrirobotic systems for various crops.
Table 8. Plant management agrirobotic systems for various crops.
CropPerceptionOperationCited Work
Apple treeForce sensorpruning[132]
HopCamera, wire detecting barstring twining[133]
GrapeThree cameraspruning[131]
Table 9. Multi-purpose agrirobotic systems for various crops.
Table 9. Multi-purpose agrirobotic systems for various crops.
Working EnvironmentPerceptionOperation(s)Cited Work
Arable cropsN/APloughing, seeding[145]
Arable cropsSonar sensor, temperatureMonitoring, spraying, fertilization, disease detection[138]
Arable cropsΤhree cameras with IR filter, humidity & ultrasonic sensorsPloughing, seeding, harvesting, spraying[146]
Arable crops, polytunnels, greenhouse2D LiDAR, ultrasonic sensors, an RGB camera & a monochromatic IR cameraHarrowing, soil sampling, phenotyping, additional tasks by combining modules[135]
Arable cropsSoil sensorsPloughing, irrigation, seeding[147]
GreenhouseTwo color cameras, machine vision sensorSpraying, weeding, additional tasks by adding or removing components[134]
Arable cropsHumidity and temperature air sensor, IR sensorsPloughing, seeding, irrigation, spraying, monitoring[136]
Urban cropsColor camera, temperature, humidity and luminosity sensorsSowing, irrigation, fumigation, pruning[139]
Arable cropsVoltage sensorsSeeding, spraying, ploughing, mowing[148]
Arable crops2D LiDAR, ultrasonic sensor, IR camera Seeding, weeding, ploughing[141]
Arable cropsN/ASowing, sprinkling, weeding, harvesting[149]
Arable cropsCloverCam, RoboWeedCamRTSeeding, weeding[137]
VineyardsN/APruning, weeding, mowing[150]
Arable cropsN/APrecision seeding, ridging discs & mechanical row crop cleaning[151]

Share and Cite

MDPI and ACS Style

Fountas, S.; Mylonas, N.; Malounas, I.; Rodias, E.; Hellmann Santos, C.; Pekkeriet, E. Agricultural Robotics for Field Operations. Sensors 2020, 20, 2672. https://doi.org/10.3390/s20092672

AMA Style

Fountas S, Mylonas N, Malounas I, Rodias E, Hellmann Santos C, Pekkeriet E. Agricultural Robotics for Field Operations. Sensors. 2020; 20(9):2672. https://doi.org/10.3390/s20092672

Chicago/Turabian Style

Fountas, Spyros, Nikos Mylonas, Ioannis Malounas, Efthymios Rodias, Christoph Hellmann Santos, and Erik Pekkeriet. 2020. "Agricultural Robotics for Field Operations" Sensors 20, no. 9: 2672. https://doi.org/10.3390/s20092672

APA Style

Fountas, S., Mylonas, N., Malounas, I., Rodias, E., Hellmann Santos, C., & Pekkeriet, E. (2020). Agricultural Robotics for Field Operations. Sensors, 20(9), 2672. https://doi.org/10.3390/s20092672

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop