This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Autonomous Navigation in Rows of Trees and High Crops with Deep Semantic Segmentation

Alessandro Navone1, Mauro Martini1, Andrea Ostuni1, Simone Angarano1 and Marcello Chiaberge1 1 Department of Electronics and Telecommunications, Politecnico di Torino, 10129, Torino, Italy. {firstname.lastname}@polito.it    Alessandro Navone
Department of Electronics and Telecommunications
Politecnico di Torino
Torino, TO, 10129
alessandro.navone@polito.it
&Mauro Martini
Department of Electronics and Telecommunications
Politecnico di Torino
Torino, TO, 10129
mauro.martini@polito.it
&Andrea Ostuni
Department of Electronics and Telecommunications
Politecnico di Torino
Torino, TO, 10129
andrea.ostuni@polito.it
&Simone Angarano
Department of Electronics and Telecommunications
Politecnico di Torino
Torino, TO, 10129
simone.angarano@polito.it
&Marcello Chiaberge
Department of Electronics and Telecommunications
Politecnico di Torino
Torino, TO, 10129
marcello.chiaberge@polito.it
Abstract

Segmentation-based autonomous navigation has recently been proposed as a promising methodology to guide robotic platforms through crop rows without requiring precise GPS localization. However, existing methods are limited to scenarios where the centre of the row can be identified thanks to the sharp distinction between the plants and the sky. However, GPS signal obstruction mainly occurs in the case of tall, dense vegetation, such as high tree rows and orchards. In this work, we extend the segmentation-based robotic guidance to those scenarios where canopies and branches occlude the sky and hinder the usage of GPS and previous methods, increasing the overall robustness and adaptability of the control algorithm. Extensive experimentation on several realistic simulated tree fields and vineyards demonstrates the competitive advantages of the proposed solution.

Keywords Mobile Robots  \cdot Precision Agriculture  \cdot Autonomous Navigation

1 Introduction

In recent years, precision agriculture has pushed the boundaries of technology to optimize crop production, improve the efficiency of farming operations, and reduce waste [1]. Modern farming systems must be able to extract synthetic key information from the environment, take or suggest optimal decisions based on that information, and execute them with high precision and timing. Deep learning techniques have shown great potential in realizing these systems by analyzing data from multiple sources, allowing for large-scale, high-resolution monitoring, and providing detailed insights for both human and robotic agents. The most recent advancements in deep learning also provide competitive advantages for real-world applications, such as model optimization for fast inference on low-power embedded hardware [2, 3] and generalization to unseen data [4, 5, 6]. At the same time, progress in service robotics has enabled autonomous mobile agents to embody AI perception systems and work in synergy with them to accomplish complex tasks in unstructured environments [7].

In particular, row-based crops are among the most studied applications (they constitute more than 75% of all planted acres of cropland across the USA [8]). In this scenario, research spans localization[9], path planning [10], navigation [11], monitoring[12], harvesting [13], spraying [14, 15], and vegetative assessment [16, 17]. A particularly challenging situation occurs when standard localization methods, like GPS, fail to reach the desired precision due to unfavorable weather conditions or line-of-sight obstruction. That is the case, for example, of dense tree canopies, as shown in a simulated pear orchard in Figure 1.

Refer to caption
Figure 1: The proposed SegMin and SegMinD algorithms allow to precisely guide an autonomous mobile robot through a dense tree row solely using an RGB-D camera. A pear crop row in Gazebo is shown in the picture.

Previous works have proposed position-agnostic vision-based navigation algorithms for row-based crops. A first vision-based approach was proposed in [18] using mean-shift clustering and the Hough transform to segment RGB images and generate the optimal central path. Later, [19] achieved promising results using multispectral images and simply thresholding and filtering on the green channel. Recently, deep-learning approaches have been successfully applied to the task. [20] proposed a classification-based approach in which a model predicts the discrete action to perform. In contrast, [21] proposed combining a segmentation model and a proportional controller to align the robot to the center of the row. Finally, a different approach was tested in [11] with an end-to-end controller based on deep reinforcement learning. Although these systems proved effective in their testing scenarios, they have only been applied in simple crops where a full view of the sky favors both GPS receivers [22] and vision-based algorithms [23].

This work tackles a more challenging scenario in which dense canopies partially or totally cover the sky, and the GPS signal is very weak. We design a navigation algorithm based on semantic segmentation that exploits visual perception to estimate the center of the crop row and align the robot trajectory to it. The segmentation masks are predicted by a deep learning model designed for real-time efficiency and trained on realistic synthetic images. The proposed navigation algorithm improves on previous works being adaptive to different terrains and crops, including dense canopies. We conduct extensive experimentation in simulated environments for multiple crops. We compare our solution with previous state-of-the-art methodologies, demonstrating that the proposed navigation system is effective and adaptive to numerous scenarios.

Refer to caption
Figure 2: Scheme of the overall proposed navigation pipeline. The RGB image is fed into the segmentation network, thus the predicted segmentation mask X^segt\hat{\textbf{X}}_{seg}^{t} is refined using the depth frame to obtain X^segDeptht\hat{\textbf{X}}_{segDepth}^{t}. The blue arrow refers to the SegMin variant, and red arrows refer to the SegMinD variant to compute the sum histogram over the mask columns. Images are taken from navigation in the tall trees simulation world.

The main contributions of this work can be summarized as follows:

  • we present two variants of a novel approach for segmentation-based autonomous navigation in tall crops, designed to tackle challenging and previously uncovered scenarios;

  • we test the resulting guidance algorithm on previously unseen plant rows scenarios such as high trees and pergola vineyards.

  • we compare the new method with state-of-the-art solutions on straight and curved vineyards, demonstrating an enhanced general and robust behaviour.

The next sections are organized as follows: Section 2 presents the proposed deep-learning-based control system for vision-based position-agnostic autonomous navigation in row-based crops, from the segmentation model to the controller. Section 3 describes the experimental setting and reports the main results for validating the proposed solution divided by sub-system. Finally, Section 4 draws conclusive comments on the work and suggests interesting future directions.

2 Methodology

This work proposes a real-time control algorithm with two variants to navigate high-vegetation orchards and arboriculture fields and improve the approach presented in [21]. The proposed system avoids exploiting the GPS signal, which can lack accuracy due to signal reflection and mitigation due to vegetation.

The working principle of the proposed control algorithms is straightforward and exploits only the RGB-D data. Both the proposed solutions consist of four main steps:

  1. 1.

    Semantic segmentation of the input RGB frame.

  2. 2.

    Processing of the output segmentation mask using depth frame data.

  3. 3.

    Searching for the direction which leads the mobile platform towards the end of the row.

  4. 4.

    Generating linear and angular velocity commands to input the mobile robot.

Nonetheless, the two proposed methods differ only for steps 2 and 3 in employing the depth frame data and in the generation of the path which the robot should follow. In contrast, the segmentation technique 1 and the command generation 4 are carried out similarly. A schematic representation of the proposed pipeline is described in Figure 2.

As in [21] a first step, an RGB frame 𝐗rgbth×w×c\mathbf{X}_{rgb}^{t}\in\mathbb{R}^{h\times w\times c} and a depth map 𝐗dth×w\mathbf{X}_{d}^{t}\in\mathbb{R}^{h\times w} are acquired by a camera placed on the front of the mobile platform at each instant tt, where hh and ww are the width of the frame and cc is the number of channels. The received RGB data is then fed to a segmentation neural network model HsegH_{seg}, which outputs a binary segmentation mask bringing the semantic information of the input frame.

X^segt=H(Xrgbt)\hat{\textbf{X}}_{seg}^{t}=H\left(\textbf{X}_{rgb}^{t}\right) (1)

where X^segt\hat{\textbf{X}}_{seg}^{t} is the estimated segmentation mask. Moreover, the segmentation masks of the last NN time instants {tN,,t}\{t-N,\dots,t\} are fused to obtain more robust information.

X^CumSegt=j=tNtX^segt\hat{\textbf{X}}_{CumSeg}^{t}=\bigcup_{j=t-N}^{t}\hat{\textbf{X}}_{seg}^{t} (2)

where X^CumSegt\hat{\textbf{X}}_{CumSeg}^{t} is the cumulative segmentation mask and the operator \bigcup represent the logical bitwise OROR operation over the last NN binary frames.

Additionally, the depth map 𝐗dt\mathbf{X}_{d}^{t} is now used to consider the segmented regions between the camera position and a given depth threshold dthd_{th} to remove useless information given by far vegetation, which is irrelevant to control the robot’s movement.

X^segDepthi=0,,hj=0,,wt(i,j)={0, if X^CumSeg(i,j)t>dth1, if X^CumSeg(i,j)tdth\hat{\textbf{X}}_{segDepth\begin{subarray}{c}i=0,\dots,h\\ j=0,\dots,w\end{subarray}}^{t}(i,j)=\begin{cases}0,\textrm{ if }\hat{\textbf{X}}_{CumSeg(i,j)}^{t}>d_{th}\\ 1,\textrm{ if }\hat{\textbf{X}}_{CumSeg(i,j)}^{t}\leq d_{th}\end{cases} (3)

where X^segDepth\hat{\textbf{X}}_{segDepth} is the resulting intersection between the cumulative segmentation frame and the depth map cut at a distance threshold dthd_{th}.

Henceforth the proposed algorithm forks in two variants, SegMin and SegMinD, respectively described in 2.1 and 2.2.

2.1 SegMin

The first variant improves the approach proposed in [21]. After processing the segmentation mask, a sum over the column is performed to obtain a histogram hw\textbf{h}\in\mathbb{R}^{w}, quantifying how much vegetation is present on each column. Hereafter, a moving average on a window of nn elements is performed over the array to smooth the values and make the control more robust to punctual noise derived from the previous passages. Ideally, the minimum of this histogram xhx_{h} corresponds to the regions where less vegetation is present and, therefore, identifies the desired central path inside the crop row. If more global minimum points are present (i.e., there is a region where no vegetation is detected), the mean of the considered points is considered to be the global minimum and, in consequence, the continuation of the row.

2.2 SegMinD

The second proposed approach consists of a variant of the previous algorithm, devised for wide rows with tall and thick canopies, which in the previous case would generate an ambiguous global minimum due to the constant presence of vegetation above the robot. This variant multiplies the previously processed segmentation mask for the normalized inverted depth datum.

X^depthInvt=X^segDeptht(1Xdtdth)\hat{\textbf{X}}_{depthInv}^{t}=\hat{\textbf{X}}_{segDepth}^{t}\bigcap\left(1-\dfrac{\textbf{X}_{d}^{t}}{d_{th}}\right) (4)

where \bigcap represents the element-wise multiplication between the binary mask X^segDeptht\hat{\textbf{X}}_{segDepth}^{t} and the inverted depth frame X^segDeptht\hat{\textbf{X}}_{segDepth}^{t} normalized over the depth threshold dthd_{th}. As in the previous case, the sum over the column is performed to obtain the 1D array h and, later on, the smoothing through a moving average. The introduced modification allows the closer elements to exert a greater influence on identifying the row direction.

2.3 Segmentation Network

We adopt the same network used in previous works on real-time crop segmentation [21, 6]. The model consists of a MobilenetV3 backbone for feature extraction and an efficient LR-ASPP segmentation head [24]. In particular, the LR-ASPP leverages effective modules such as depth-wise convolutions, channel-wise attention, and residual skip connections to provide an effective trade-off between accuracy and inference speed. The model is trained with a similar procedure to [6] on the AgriSeg dataset111https://pic4ser.polito.it/AgriSeg. Further details on the training strategy and hyperparameters are provided in Section 3.

2.4 Robot heading control

The objective of the controller pipeline consists in keeping the mobile platform at the center of the row, which, in this work, is considered equivalent to keeping the row center in the middle of the camera frame. Therefore, as defined in the previous step, the minimum of the histogram should be centered in the frame width. The distance dd from the center of the frame and the minimum is defined as:

d=xhw2d=x_{h}-\dfrac{w}{2} (5)

The linear and angular velocities are then generated through custom functions as in [25].

vx=vx,max(1d2w22)v_{x}=v_{x,max}\left(1-\frac{d^{2}}{\frac{w}{2}^{2}}\right) (6)
ωz=ωz,gaind\omega_{z}=-\omega_{z,gain}\cdot d (7)

where vx,maxv_{x,max} is the maximum achievable linear speed and ωz,gain\omega_{z,gain} is the andular gain. In order to avoid abrupt changes in the robot’s motion, the final velocities v¯x\bar{v}_{x} and ω¯z\bar{\omega}_{z} commands are smoothed with an Exponential Moving Average (EMA) as:

[v¯xtω¯zt]=(1λ)[v¯xt1ω¯zt1]+λ[vxtωzt]\begin{bmatrix}\bar{v}_{x}^{t}\\ \bar{\omega}_{z}^{t}\end{bmatrix}=(1-\lambda)\begin{bmatrix}\bar{v}_{x}^{t-1}\\ \bar{\omega}_{z}^{t-1}\end{bmatrix}+\lambda\begin{bmatrix}{v}_{x}^{t}\\ \omega_{z}^{t}\end{bmatrix} (8)

where tt is the time step and λ\lambda is a chosen weight.

3 Experiments and Results

3.1 Simulation Environment

The proposed control algorithm was tested through the use of Gazebo222https://gazebosim.org simulation software. The software was selected because of its compatibility with ROS 2 and can incorporate plugins that simulate sensors, such as cameras. A Clearpath Jackal model was utilized to assess the algorithm’s effectiveness. The URDF file, available through Clearpath Robotics, contains all the necessary information regarding the mechanical structure and joints of the robot. During the simulation, an Intel Realsense D435i plugin was utilized, positioned 20 cm in front of the robot’s center, and tilted 1515^{\circ} upwards. This positioning gave the camera a better view of the upper branches of trees.

The navigation algorithm was tested in four different custom simulation environments: a common vineyard, a pergola vineyard characterized by vine poles and shoots above the row, a pear field constituted by small size trees, and a high trees field where canopies of the trees are merged above the row. Each simulated field adopts a different terrain, miming the irregularity of uneven terrain. The detailed measurements of the simulation world are described in Table 1.

During the experimental part of this work, we consider frame dimensions equal to (h,w)=(224,224)(h,w)=(224,224), which is the same size as the input and the output of the neural network model, with the number of channels c=3c=3. The maximum linear velocity has been fixed to vx,max=0.5m/sv_{x,max}=0.5m/s, and the maximum angular velocity has been fixed to ωz,max=1rad/s\omega_{z,max}=1rad/s. The angular velocity gain ωz,gain\omega_{z,gain} has been fixed to 0.010.01 and the EMA buffer size has been fixed to 3. The depth threshold has been changed according to the various crops. In particular, it has been fixed to 5 m in the case of vineyards, while it was increased to 8 m for pear trees and pergola vineyards and 10 m for tall trees.

RGB Masked Depth Histogram
(a) Refer to caption Refer to caption Refer to caption
(b) Refer to caption Refer to caption Refer to caption
(c) Refer to caption Refer to caption Refer to caption
(d) Refer to caption Refer to caption Refer to caption
Figure 3: Sample outputs of the proposed SegMinD algorithm for High Trees (a), Pear Trees (b), Pergola Vineyard (c), and Vineyard (d). Predicted segmentation masks are refined cutting values exceeding a depth threshold. The sum over mask columns provide the histograms used to identify the centre of the row as its global minimum.
Table 1: Size of the different simulated crops, referring to the average values of the distance between rows, the distance between plants on the row and the heights of the plants.
Gazebo worlds Rows distance [m] Plant distance [m] Height [m]
Common vineyard 1.8 1.3 2.0
Pergola vineyard 6.0 1.5 2.9
Pear field 2.0 1.0 2.9
High trees field 7.0 5.0 12.5
Refer to caption Refer to caption
(a) (b)
Refer to caption Refer to caption
(c) (d)
Figure 4: Gazebo simulated environments used to test the SegMin approach in relevant different crops rows: wide rows composed of high trees (a), a narrow pear tree row (b), a pergola vineyard with asymmetric rows (c), straight and curved vineyard rows (d). In the last case, the tests were carried out in the second row from above and the second row from below.

3.2 Segmentation Network Training and Evaluation

We train the crop segmentation model using a subset of the AgriSeg segmentation dataset [6]. In particular, for the High Tree and Pear crops, we train on Generic Tree splits 1 and 2, and on Pear; for Vineyards, we train on Vineyard and Pergola Vineyard (note that the testing environments are different from the ones from which the training samples are generated). In both cases, the model is trained for 50 epochs with Adam optimizer and learning rate 3×1043\times 10^{-4}. We apply data augmentation by randomly applying cropping, flipping, greyscaling, and random jitter to the images. Our experimentation code is developed in Python 3 using TensorFlow as the deep learning framework. We train models starting from ImageNet pretrained weights, so the input size is fixed to (224 × 224). The All the training runs are performed on a single Nvidia RTX 3090 graphic card.

3.3 Navigation Results

Table 2: Navigation results obtained in different test fields with the SegMin, SegMinD and previous work SegZeros segmentation-based algorithms. The metrics test the effectiveness of the navigation (clearance time) and its precision with Mean Absolute Error (MAE) and Mean Squared Error (MSE) between obtained and ground truth path. The cumulative heading average γ[rad]\gamma[rad], the mean linear velocity vavg[m/s]v_{avg}[m/s] and the standard deviation of the angular velocity ωstddev[rad/s]\omega_{stddev}[rad/s] commands provide relevant kinematic information of the robot while navigating.
Test Field Method Clearance time [s] MAE [m] MSE [m] Cum. γavg\boldsymbol{\gamma_{avg}} [rad] 𝒗𝒂𝒗𝒈[𝒎/𝒔]\boldsymbol{v_{avg}[m/s]} 𝝎𝒔𝒕𝒅𝒅𝒆𝒗[𝒓𝒂𝒅/𝒔]\boldsymbol{\omega_{stddev}[rad/s]}
High Trees SegMin 40.409 ± 0.117 0.265 ± 0.005 0.084 ± 0.003 0.079 ± 0.001 0.487 ± 0.000 0.054 ± 0.002
SegMinD 40.440 ± 0.515 0.174 ± 0.006 0.036 ± 0.002 0.048 ± 0.002 0.484 ± 0.006 0.063 ± 0.019
Pear Trees SegMin 42.058 ± 1.228 0.034 ± 0.012 0.002 ± 0.001 0.013 ± 0.002 0.483 ± 0.003 0.108 ± 0.054
SegMinD 42.259 ± 1.912 0.031 ± 0.017 0.002 ± 0.002 0.016 ± 0.004 0.477 ± 0.009 0.026 ± 0.004
Pergola Vineyard SegMin 40.859 ± 0.386 0.077 ± 0.011 0.011 ± 0.003 0.030 ± 0.022 0.479 ± 0.003 0.174 ± 0.021
SegMinD 41.135 ± 0.329 0.097 ± 0.052 0.015 ± 0.014 0.029 ± 0.011 0.475 ± 0.004 0.204 ± 0.032
Straight Vineyard SegMin 50.509 ± 0.305 0.105 ± 0.003 0.014 ± 0.001 0.033 ± 0.002 0.487 ± 0.000 0.079 ± 0.011
SegMinD 50.629 ± 0.282 0.110 ± 0.005 0.018 ± 0.003 0.026 ± 0.009 0.486 ± 0.001 0.088 ± 0.005
SegZeros 53.695 ± 1.029 0.138 ± 0.025 0.024 ± 0.010 0.027 ± 0.004 0.457 ± 0.008 0.089 ± 0.008
Curved Vineyard SegMin 53.321 ± 0.249 0.115 ± 0.008 0.017 ± 0.002 0.036 ± 0.008 0.487 ± 0.001 0.088 ± 0.021
SegMinD 51.444 ± 1.030 0.093 ± 0.005 0.012 ± 0.001 0.015 ± 0.004 0.484 ± 0.007 0.065 ± 0.008
SegZeros 71.048 ± 27.132 0.108 ± 0.044 0.019 ± 0.009 0.045 ± 0.008 0.395 ± 0.127 0.114 ± 0.039

The overall navigation pipeline of SegMin and its variant SegMinD are tested in realistic crops fields in simulation using relevant metrics for visual-based control without precise localization of the robot, as done in previous works [21, 11]. The camera frames are published at a frequency of 30 Hz, while the inference is carried out at 20 Hz, and the controllers publish the velocity commands at 5 Hz.The evaluation has been performed using the testing package of the open-source PIC4rl-gym333https://github.com/PIC4SeR/PIC4rl_gym in Gazebo [26]. The selected metrics aims at evaluating the effectiveness of the navigation (clearance time) as well as the precision, quantitatively comparing the obtained trajectories with a ground truth one though Mean Absolute Error (MAE) and Mean Squared Error (MSE). The ground truth trajectories have been computed averaging the curve obtained interpolating the plants poses in the rows. For the asymmetric pergola vineyard case, the row is intended as the portion of the pergola without vegetation on top, as shown in Figure 4 (c). The response of the algorithms to terrain irregularity and rows geometry is also studied including in the test significant kinematic information of the robot. The cumulative heading average γ[rad]\gamma[rad] along the path is considered, together with the mean linear velocity vavg[m/s]v_{avg}[m/s] and the standard deviation of the angular velocity ωstddev[rad/s]\omega_{stddev}[rad/s] commands predicted to keep the robot correctly oriented. The mean value of ω\omega is always close to zero due to the consecutive correction of the robot orientation.

The complete results collection is reported in Table 2. For each metric, an average value and the standard deviation are indicated, since all the experiments have been repeated over 3 runs on a 20 m long track in each crop row. The proposed method demonstrates to solve the problem of guiding the robot through trees rows with thick canopies (high trees and pears) without a localization system, as well as in peculiar scenarios such as the pergola vineyards. The identification of plants branches and wooden supports hinder the usage of previously existing segmentation-based solutions, that were based on the assumption of finding a free passage solely considering the zeros of the binary segmentation mask [21]. We refer to this previous method as SegZeros in the results comparison, that we tested using the same segmentation neural network.

The SegMin approach based on histogram minimum search demonstrate to be a robust solution to guide the robot through trees rows. The introduction of the depth inverse values as weighting function allows SegMinD to be further increase the precision of the algorithm in following the central trajectory of the row in complex cases such as wide rows (high trees) and curved rows (curved vineyard). The different sum histograms obtained with SegMin and SegMinD are directly compared in Figure 5, showing the sharper trend and the global minimum isolation obtained including the depth values. Moreover, the novel methods show competitive performance also with standard crops rows where a free passage to the end of the row can be seen in the mask without the disturbance of canopies. The histogram minimum approach significantly reduce the navigation time and the trajectory precision in vineyard rows (straight and curved) compared to previous segmentation-based baseline method. The search of plant-free zero clusters in the map results to be less robust and efficient, leading the robot to undesired stops during the navigation, and to an overall slower and more oscillating behaviour.

RGB SegMin SegMinD
Refer to caption Refer to caption Refer to caption
Figure 5: Comparison of the two histograms obtained using the two different algorithms, given the RGB frame on the right. It can be noticed how SegMinD offers a narrower and less ambiguous global minimum point.

Nonetheless, the trajectories obtained with the SegMin, SegMinD and SegZeros algorithms are also visually shown in Figure 6 inside representative scenarios: a cluttered, narrow row with small pear trees, a wide row with high trees, and curved vineyards with state-of-the-art method SegZeros.

Refer to caption
Refer to caption
Refer to caption
Figure 6: Trajectories comparison between our proposed algorithms (SegMin and SegMinD) and the ground truth central path (GT): Pears (top), High Trees (center), Curved Vineyard (bottom). In the last graph, the trajectory generated with the SegZeros algorithm is also reported for comparison.

4 Conclusions

In this work, we presented a novel method to guide to a service autonomous platform through crops rows where a precise localization signal is often occluded by the vegetation. Trees rows represented an open problem in row crops navigation, since previous works based on image segmentation or processing fail due to the presence of branches and canopies covering the free passage for the rover in the image. The proposed pipeline SegMin and SegMinD overcome this limitation introducing a global minimum search on the sum histogram over the mask columns. The experiments conducted demonstrate the ability to solve the navigation task in wide and narrow trees rows and, nonetheless, the improvement in efficiency and robustness provided by our method over previous works in generic vineyards scenarios.

Future works will see the test of the overall system in real-world trees rows, orchards and vineyards to further validate the robustness of the solution with respect to sim2real gap problems and hardware resources. A successful outcome is expected according to the efficient architecture adopted for the segmentation neural network, thought for low-power system applications, and the realistic features of training data and simulation.

Acknowledgements

This work has been developed with the contribution of Politecnico di Torino Interdepartmental Centre for Service Robotics PIC4SeR444www.pic4ser.polito.it.

References

  • [1] Zhaoyu Zhai, José Fernán Martínez, Victoria Beltran, and Néstor Lucas Martínez. Decision support systems for agriculture 4.0: Survey and challenges. Computers and Electronics in Agriculture, 170:105256, 2020.
  • [2] Vittorio Mazzia, Aleem Khaliq, Francesco Salvetti, and Marcello Chiaberge. Real-time apple detection system using embedded systems with hardware accelerators: An edge ai application. IEEE Access, 8:9102–9114, 2020.
  • [3] Simone Angarano, Francesco Salvetti, Vittorio Mazzia, Giovanni Fantin, Dario Gandini, and Marcello Chiaberge. Ultra-low-power range error mitigation for ultra-wideband precise localization. In Intelligent Computing: Proceedings of the 2022 Computing Conference, Volume 2, pages 814–824. Springer, 2022.
  • [4] Mauro Martini, Vittorio Mazzia, Aleem Khaliq, and Marcello Chiaberge. Domain-adversarial training of self-attention-based networks for land cover classification using multi-temporal sentinel-2 satellite imagery. Remote Sensing, 13(13):2564, 2021.
  • [5] Simone Angarano, Mauro Martini, Francesco Salvetti, Vittorio Mazzia, and Marcello Chiaberge. Back-to-bones: Rediscovering the role of backbones in domain generalization. arXiv preprint arXiv:2209.01121, 2022.
  • [6] Simone Angarano, Mauro Martini, Alessandro Navone, and Marcello Chiaberge. Domain generalization for crop segmentation with knowledge distillation. arXiv preprint arXiv:2304.01029, 2023.
  • [7] Andrea Eirale, Mauro Martini, Luigi Tagliavini, Dario Gandini, Marcello Chiaberge, and Giuseppe Quaglia. Marvin: An innovative omni-directional robotic assistant for domestic environments. Sensors, 22(14):5261, 2022.
  • [8] Daniel Bigelow and Allison Borchers. Major uses of land in the united states, 2012. Economic Information Bulletin Number 178, (1476-2017-4340):69, 2017.
  • [9] Wera Winterhalter, Freya Fleckenstein, Christian Dornhege, and Wolfram Burgard. Localization for precision navigation in agricultural fields—beyond crop row following. Journal of Field Robotics, 38(3):429–451, 2021.
  • [10] Francesco Salvetti, Simone Angarano, Mauro Martini, Simone Cerrato, and Marcello Chiaberge. Waypoint generation in row-based crops with deep learning and contrastive clustering. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2022, Grenoble, France, September 19–23, 2022, Proceedings, Part VI, pages 203–218. Springer, 2023.
  • [11] Mauro Martini, Simone Cerrato, Francesco Salvetti, Simone Angarano, and Marcello Chiaberge. Position-agnostic autonomous navigation in vineyards with deep reinforcement learning. In 2022 IEEE 18th International Conference on Automation Science and Engineering (CASE), pages 477–484, 2022.
  • [12] Lorenzo Comba, Alessandro Biglia, Davide Ricauda Aimonino, Paolo Barge, Cristina Tortia, and Paolo Gay. 2d and 3d data fusion for crop monitoring in precision agriculture. In 2019 IEEE international workshop on metrology for agriculture and forestry (MetroAgriFor), pages 62–67. IEEE, 2019.
  • [13] C Wouter Bac, Eldert J van Henten, Jochen Hemming, and Yael Edan. Harvesting robots for high-value crops: State-of-the-art review and challenges ahead. Journal of Field Robotics, 31(6):888–911, 2014.
  • [14] Deepak Deshmukh, Dilip Kumar Pratihar, Alok Kanti Deb, Hena Ray, and Nabarun Bhattacharyya. Design and development of intelligent pesticide spraying system for agricultural robot. In Hybrid Intelligent Systems: 20th International Conference on Hybrid Intelligent Systems (HIS 2020), December 14-16, 2020, pages 157–170. Springer, 2021.
  • [15] Ron Berenstein, Ohad Ben Shahar, Amir Shapiro, and Yael Edan. Grape clusters and foliage detection algorithms for autonomous selective vineyard sprayer. Intelligent Service Robotics, 3(4):233–243, 2010.
  • [16] GuoSheng Zhang, TongYu Xu, YouWen Tian, Han Xu, JiaYu Song, and Yubin Lan. Assessment of rice leaf blast severity using hyperspectral imaging during late vegetative growth. Australasian Plant Pathology, 49:571–578, 2020.
  • [17] Aijing Feng, Jianfeng Zhou, Earl D Vories, Kenneth A Sudduth, and Meina Zhang. Yield estimation in cotton using uav-based multi-sensor imagery. Biosystems Engineering, 193:101–114, 2020.
  • [18] Mostafa Sharifi and XiaoQi Chen. A novel vision based row guidance approach for navigation of agricultural mobile robots in orchards. In 2015 6th International Conference on Automation, Robotics and Applications (ICARA), pages 251–255, 2015.
  • [19] Josiah Radcliffe, Julie Cox, and Duke M. Bulanon. Machine vision for orchard navigation. Computers in Industry, 98:165–171, 2018.
  • [20] Peichen Huang, Lixue Zhu, Zhigang Zhang, and Chenyu Yang. An end-to-end learning-based row-following system for an agricultural robot in structured apple orchards. Mathematical Problems in Engineering, 2021, 2021.
  • [21] Diego Aghi, Simone Cerrato, Vittorio Mazzia, and Marcello Chiaberge. Deep semantic segmentation at the edge for autonomous navigation in vineyard rows. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3421–3428. IEEE, 2021.
  • [22] Md. Shaha Nur Kabir, Ming-Zhang Song, Nam-Seok Sung, Sun-Ok Chung, Yong-Joo Kim, Noboru Noguchi, and Soon-Jung Hong. Performance comparison of single and multi-gnss receivers under agricultural fields in korea. Engineering in Agriculture, Environment and Food, 9(1):27–35, 2016.
  • [23] Shahzad Zaman, Lorenzo Comba, Alessandro Biglia, Davide Ricauda Aimonino, Paolo Barge, and Paolo Gay. Cost-effective visual odometry system for vehicle motion control in agricultural environments. Computers and Electronics in Agriculture, 162:82–94, 2019.
  • [24] Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF international conference on computer vision, pages 1314–1324, 2019.
  • [25] Simone Cerrato, Vittorio Mazzia, Francesco Salvetti, and Marcello Chiaberge. A deep learning driven algorithmic pipeline for autonomous navigation in row-based crops. CoRR, abs/2112.03816, 2021.
  • [26] Mauro Martini, Andrea Eirale, Simone Cerrato, and Marcello Chiaberge. Pic4rl-gym: a ros2 modular framework for robots autonomous navigation with deep reinforcement learning. arXiv preprint arXiv:2211.10714, 2022.