Access to this full-text is provided by Springer Nature.
Content available from Chemical and Biological Technologies in Agriculture
This content is subject to copyright. Terms and conditions apply.
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
https://doi.org/10.1186/s40538-021-00217-8
REVIEW
Drone andsensor technology forsustainable
weed management: areview
Marco Esposito†, Mariano Crimaldi†, Valerio Cirillo* , Fabrizio Sarghini and Albino Maggio
Abstract
Weeds are amongst the most impacting abiotic factors in agriculture, causing important yield loss worldwide. Inte-
grated Weed Management coupled with the use of Unmanned Aerial Vehicles (drones), allows for Site-Specific Weed
Management, which is a highly efficient methodology as well as beneficial to the environment. The identification of
weed patches in a cultivated field can be achieved by combining image acquisition by drones and further process-
ing by machine learning techniques. Specific algorithms can be trained to manage weeds removal by Autonomous
Weeding Robot systems via herbicide spray or mechanical procedures. However, scientific and technical under-
standing of the specific goals and available technology is necessary to rapidly advance in this field. In this review, we
provide an overview of precision weed control with a focus on the potential and practical use of the most advanced
sensors available in the market. Much effort is needed to fully understand weed population dynamics and their com-
petition with crops so as to implement this approach in real agricultural contexts.
Keywords: Site-specific weed management, UAVs, Precision agriculture, Weed detection, Crop–weed interaction
© The Author(s) 2021. This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and
the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material
in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material
is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the
permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creat iveco
mmons .org/licen ses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creat iveco mmons .org/publi cdoma in/
zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
Introduction
Biotic threats such as insects, weeds, fungi, viruses, and
bacteria can broadly affect crop yield and quality. Among
these, weeds are the most impacting problem causing
remarkable yield loss worldwide [1]. e most char-
acterized effect of weeds is competition for resources
such as light [2], water [3], space [4], and nutrients [5].
In addition, specific chemical signals and/or toxic mol-
ecules produced by weeds may interfere with a normal
crop development [6]. A distinctive trait of wild species,
including weeds, is their high physiological, morphologi-
cal, and anatomical plasticity which makes them more
tolerant than crop species to environmental stressors [7–
10]. Moreover, weeds interact with other biological com-
ponents of the environment, acting as refuge for plant
pests such as insects, fungi, and bacteria that can harm
close in crops [11–13]. For example, wild oats (Avena
fatua L.) can harbor the etiological agents of the powdery
mildew in crops such as wheat (Triticum aestivum L.),
oats, and barley (Hordeum vulgare L.) [14]; altamisa (Par-
thenium hysterophorus L.) can be a secondary host of the
common hairy caterpillar (Diacrisia obliqua Walk.) [15,
16]; Cyperus rotundus can host the root-knot Meloido-
gyne graminicola and, therefore, can contribute to their
spreading in the field [17]. Finally, weed infestation may
affect fresh and processed products quality such as beer,
wine, forage [18, 19]. In this respect, weed residuals may
cause accumulation of off-flavors products [20, 21], or in
some cases, can make them harmful to humans and ani-
mals [22, 23]. Weeds may also contain high levels of aller-
gens and/or toxic metabolites that, if ingested, can cause
asthma, skin rash, and other reactions [24, 25].
Most weed research aims at developing strategies that
can reduce the deleterious impact of the interspecific
competition between crops and weeds and recent tech-
nological advances may further contribute to this scope,
while improving the sustainability of weed control [26–
28]. Worldwide, weed competition causes severe yield
reduction in all major crops, such as wheat (23%), soy-
bean (37%), rice (37%), maize (40%), cotton (36%), and
potato (30%) [1]. Yearly, weeds cause 50% yield losses of
Open Access
*Correspondence: valerio.cirillo@unina.it
†Marco Esposito and Mariano Crimaldi contributed equally to this work
Department of Agricultural Sciences, University of Naples Federico II, Via
Università 100, 80055 Portici, NA, Italy
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 2 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
corn and soybean productivity in North America. For
corn, this equates to a loss of 148 million tons for an
economical loss of over $26.7 billion [29]. In Australia,
yield loss due to weeds accounts for 2.76 million tons of
grain from different plants, including wheat, barley, oats,
canola, sorghum, and pulses [30]. e annual global eco-
nomic loss caused by weeds has been estimated to be
more than $100 billion U.S. dollars [31], despite world-
wide annual herbicide sales in the range of $25 billion
[32]. In Europe, herbicides are the second most-sold pes-
ticides. ey accounted for 35% of all pesticide sales in
2018, overcoming insecticides and acaricides (Fig.1) [33].
Weed management requires anintegrated approach
In 2050, the world population will quadruplicate, reach-
ing 9.15 billion people [34]. However, the predicted
increase in food demand will be hardly met by the cur-
rent production system [35]. Also, climate change will be
an additional challenge for the human food supply in the
near future [36]. Among all the processes affecting crop
productivity, weed management will be one of the hard-
est challenges [37]. Mechanical and chemical weed con-
trol has disadvantages that probably will impede them
to be effective for future weed management [38–43].
Mechanical methods are scarcely efficient, and herbi-
cides have a high ecological impact. An approach that
minimizes the drawbacks of mechanical and chemical
weed control is Integrated Weed Management (IWM).
IWM combines chemical, biological, mechanical, and/
or crop management methods, and represents a model
to improve the efficiency and sustainability of weed con-
trol [3, 44]. In contrast to traditional methods, IWM
integrates several agro-ecological aspects such as the role
of conservation tillage and crop rotation on weeds seed
bank dynamics [10], the ability to forecast the critical
period of weed interference and their competition with
crops [45, 46], and the specific critical levels of crops/
weeds interaction [47]. erefore, an effective IWM must
rely on a thorough knowledge of crop-weeds competition
dynamics, which currently represents one of the most
active research areas in weed science [48, 49].
New technologies forsite‑specic weed management
Precision agriculture relies on technologies that combine
sensors, information systems, and informed management
to optimize crop productivity and to reduce the environ-
mental impact [50]. Nowadays, precision agriculture has
a broad range of applications and it is employed in dif-
ferent agricultural contexts including pests control [51],
fertilization, irrigation [52, 53], sowing [54] and harvest-
ing [55]. Precision agriculture can be effectively applied
to IWM also. In the last decade, precision agriculture
has rapidly advanced because of technological innova-
tions in the areas of sensors [56], computer hardware
[57], nanotechnology [58], unmanned vehicles systems
and robots [59] that may allow for specific identification
of weeds that are present in the field [47]. Unmanned
aerial vehicles (UAV) are one of the most successful tech-
nologies applied in precision agriculture [60]. Unmanned
Vehicles systems are mobile Aerial (UAV) or Terrestrial
(UTV) platforms that provide numerous advantages for
the execution and monitoring of farming activities [61].
UAVs can be highly valuable since they allow for Site-
Specific Weed Management (SSWM) (Fig.2). SSWM is
Fig. 1 Percentage (of total volume in kilograms) of pesticide sales by category in Europe in 2018 [33]
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 3 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
an improved weed management approach for highly effi-
cient and environmentally safe control of weed popula-
tions [28], enabling precise and continuous monitoring
and mapping of weed infestation. SSWM consents to
optimize weed treatments for each specific agronomical
situation [62]. e combination of UAVs with advanced
cameras and sensors, able to discern specific weeds [63],
and GPS technologies, that provide geographical infor-
mation for field mapping, can help in precisely monitor-
ing large areas in a few minutes. anks to more accurate
planning of weed management that can increase mechan-
ical methods effectiveness and/or reduce herbicide
spread [64], the potential agro-ecological and economic
implications of SSWM are remarkable, yielding lower
production costs, reducing the onset of weed resistance,
improving biodiversity, and containing environmental
impacts [65]. e application of UAVs to weed control
can, therefore, contribute to improve the sustainability of
future agricultural production systems that must comply
with an increasing world population [34, 35].
UAVs remote sensing techniques andsensors
UAVs have become a common tool in precision agricul-
ture [66, 67]. anks to their affordability, user-friendli-
ness and versatility, UAVs are often the primary choice
for fast and precise in situ remote sensing or survey
operations. Despite their versatility, these systems may
be used for different purposes, depending on the sen-
sors they carry on. Ongoing research is looking at the
best solutions to integrate data collected from sensors on
UAVs, ground sensors and other data sources for better
management of punctual operations in the field, with a
particular focus on smart agriculture and big data man-
agement [68, 69].
Although UAVs systems do not offer the same territo-
rial coverage as satellites, they offer a spatial and temporal
resolution that other systems do not [70, 71]. From an
economic point of view, the use of drones requires the
investment to buy a UAV system with at least a 0.1cm/px
resolution RGB camera, a trained pilot for flight manage-
ment and post-processing software capabilities. e ini-
tial UAV investment is compensated by the repeatability
of flights, which increases the frequency of datasets deliv-
ered, and the higher resolution compared to other sys-
tems [72, 73]. UAVs systems also have further advantages:
(1) the possibility to collect easily deployable data in real
time (excluding post-processing); (2) they can be used to
survey areas with high level of hazard and/or difficult to
reach; (3) they allow operators to collect data even with
unfavorable weather conditions, such as in very cloudy or
foggy days, under which satellite detection systems fail or
produce very altered datasets [71]. e most important
sensors available as payload are mainly categorized into
three classes depending on the spectral length and num-
ber they can record:
• RGB (Red, Green, Blue) or VIS (Visible) sensors
• Multispectral sensors
• Hyperspectral sensors
RGB/VIS sensors
e RGB or VIS sensors are the most common and
largely available commercial cameras (Table 1). eir
possible applications have been the focus of most
research for years due to their potential and low-cost
operational requirements [74, 75].
ese sensors are used to calculate vegetation indices
such as the Green/Red Vegetation Index (GRVI), Green-
ness Index (GI) and Excessive Greenness (ExG) with
acceptable or high levels of accuracy [76, 77]. Also, RGB
sensors have been increasingly used for machine learn-
ing techniques in object recognition, phenology, pathol-
ogies, and similar purposes. e typical workflow of
processing RGB images from UAVs for remote sensing is:
1. pre-flight planning, 2. flight and image acquisition, 3.
post-processing and indexes or dataset extrapolation [71].
Phase 1 is critical and essential to collect data of useful
quality for the purpose. In the pre-flight planning phase,
the parameters to consider are the definition of the study
area, the flight altitude, site topography, weather fore-
cast and local regulations for unmanned flights. In phase
2, it is recommended to keep the data flow sufficient to
store data and to check if the acquisition platform can
acquire the amount of data required. It could be possi-
ble to encounter I/O errors due to the inadequacy of the
platform with consequent loss of information or abor-
tion of the mission. In phase 3, for RGB sensors, there is
no need to perform radiometric calibration, which is the
Fig. 2 Site-specific weed management (SSWM) scheme realized by
drones and its economical and agro-ecological implications
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 4 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
case when using multispectral and hyperspectral sensors.
RGB data can be used per se or to create a georeferenced
orthomosaic. In this case, the individual images are recti-
fied, georeferenced using GPS data and stitched together
to form a single image (orthomosaic) covering the entire
study area. Orthomosaics can be generated either with
RGB values as they are or after calculating the desired
vegetation indices [77]. If RGB images are to be used
in machine learning algorithms, the workflow is differ-
ent [78–81]. In this case, it is necessary to collect a large
dataset of images for the training and testing of the algo-
rithm [82]. is dataset may already be available from
third-party sources, such as PlantVillage [83] or PlantDoc
[84]. Alternatively, it can be created from scratch if the
purpose of the research is not covered by existing data-
sets [85]. In this case, the acquisition, selection and pro-
cessing of the images are critical, because the final dataset
can affect both the training and the use of the neural net-
work, with risks of producing biased results [86].
Multispectral sensors
e multispectral sensors are used for a wider range of
calculations of vegetation indices as they can rely on a
higher number of radiometric bands. A comparison of
the most common multispectral sensors, specific for
UAV systems, is shown in Table2.
With multispectral sensors, the range of vegetation indi-
ces that can be monitored is considerably extended com-
pared to those that can be calculated with only three RGB
bands. Moreover, the workflow has minor variations. For
these sensors, in phase 1, the radiometric calibration and
atmospheric correction phases are strictly required. Many
multispectral sensors, such as the Micasense RedEdge
series or the Parrot Sequoia + , have downwelling irradi-
ance sensors and a calibrated reflectance panel to address
some of the requirements for radiometric calibration [87].
Due to a lower resolution of the sensors compared to RGB
ones, a lower flight altitude and an adequate horizontal
and vertical overlap of recorded images must be taken into
Table 1 RGB cameras andtheir main specications
a * UAV with already supplied camera. Payload not interchangeable
Camera model Sensor type
andresolution
[Mpx]
Sensor format Sensor Size [mm] Weight [kg] Price (approx.) [€]
Canon EOS 5d Mark IV CMOS 30.4 Full Frame 36.0 × 24.0 ca. 1.0 ca. 1000
Nikon D610 CMOS 24.3 Full Frame 36.0 × 24.0 ca. 1.250 ca. 1000
Sony Alpha 7R II CMOS 42 Full Frame Mirrorless 35.0 × 24.0 ca. 0.6 ca. 1200
Sony Alpha a6300 CMOS 24 Small Frame Mirrorless 23.5 × 15.6 ca. 0.8 ca. 800
Panasonic Lumix DMC GX8 CMOS 20 Small Frame Mirrorless 17.3 × 13 ca. 0.5 ca. 1000
Panasonic Lumix DMC GX80 DLMOS 16 Small Frame Mirrorless 17.3 × 13 ca. 0.5 ca. 500
DJI Phantom 4 Pro * CMOS 20 Small Frame 13.2 × 8.8 ca. 1.5 (with UAV) ca. 1500 (with UAV)
DJI Mavic 2 ProaCMOS 20 Small Frame 13.2 × 8.8 ca. 1.5 (with UAV) ca. 1500 (with UAV)
Table 2 Multispectral sensors andtheir main specications
Camera model Resolution [Mpx] Spectral bands Ground sample distance
[cm/px]Weight [kg] Price (approx.) [€]
Micasense RedEdge-M 1280 × 960 (1.2 Mpx per EO
band) Red, Green, Blue, Near-
Infrared, Red Edge 8 (per band) at 120 m AGL ca. 0.180 ca. 5000
Micasense RedEdge-MX 1280 × 960 (1.2 Mpx per EO
band) Blue, green, red, red edge,
near infrared (NIR) 8 (per band) at 120 m AGL ca. 0.231 ca. 5000
Micasense Altum 2064 × 1544 (3.2 Mpx per EO
band) 160 × 120 thermal
infrared
EO: Blue, green, red, red
edge, near-infrared (NIR)
LWIR: thermal infrared
8–14 µm
5.2 cm per pixel (per EO
band) at 120 m AGL—
81 cm per pixel (thermal)
at 120 m
ca. 0.405 ca. 6000
TertaCam MCAW 6 1.3 6 user selectable narrow
bands (450–1000 µm) – ca. 0.550 ca. 17000
TetraCam ADC Lite 3.2 Green, Red, Near-Infrared
(NIR) – ca. 0.2 ca. 3000
TetraCam ADC Micro 3.2 Green, Red, Near-Infrared
(NIR) – ca. 0.09 ca. 3000
Parrot Sequoia + 1.2 Blue, Green, Red, Red Edge,
Near-Infrared (NIR) – ca. 0.7 ca. 5000
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 5 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
account to obtain an adequate ground resolution for the
surveyed objective and to avoid missing data [88]. In phase
2, having a higher number of radiometric bands to record,
the dataflow will be higher so is critical to avoid I/O errors,
missing data or mission failures [89]. Due to multi-lenses
nature of the sensors in phase 3, the data collected suffer
from the parallax problem. As a consequence, images have
to be rectified, georeferenced and must be stacked to gen-
erate a single image with different radiometric levels, and
calibrated with the downwelling irradiance sensors data
acquired during the flight [90]. After this procedure, it is
possible to generate a multispectral orthomosaic and then
calculate the requested indexes [91]. Multispectral images
are also used in machine learning applications [80, 85, 92]
taking into account the multi-camera nature of sensors
and the different bands recorded. anks to the availabil-
ity of a higher number of radiometric bands, the machine
learning algorithms can be extended to not-visible recog-
nition such as early stage plant disease, field quality assess-
ment, soil water content, and more [91].
Hyperspectral sensors
e hyperspectral sensors can record hundreds to thou-
sands of narrow radiometric bands, usually in visible and
infrared ranges. To deal with hyperspectral applications, the
choice of number and radiometric range of bands is criti-
cal. Each band or combination of bands, being very narrow,
can detect a specific field characteristic. Each hyperspectral
sensor can detect only a certain number of bands, so the
aim of survey must be very clear to choose the right sensor.
Although hyperspectral sensors have decreased in price in
recent years, they are still an important starting investment
since they are much more expensive than RGB and multi-
spectral sensors. In addition, they are heavier and bigger
than other sensors, often making their use on UAV systems
difficult and/or excessively onerous in terms of payload.
Some of most used hyperspectral sensors in UAVs applica-
tion and their main characteristics are shown in Table3.
In this case, the workflow for radiometric calibration is
more complex compared to other sensors. Some calibration
methods needed for these sensors are derived from manned
aircraft hyperspectral platforms, based on artificial targets
to assess data quality, to correct radiance, and to generate
a high-quality reflectance data-cube [93]. In phase 1, the
planning must also be carried out in time and not only in
space because, in addition to the spectrometric resolution,
hyperspectral sensors have a temporal resolution due to the
different acquisition method. In phase 2, it should be con-
sidered that both images’ size and data flow are bigger than
multispectral/RGB images. Moreover, these sensors may
acquire a large amount of data, but the payload limitations
of UAVs may not allow the transport of adequate file storage
systems. Phase 3 for hyperspectral images is critical: qual-
ity assessment is one of the critical issues of hyperspectral
data and some problems associated with the quality of the
images have not been completely overcome. Among those,
the stability of the sensor itself (due to the nature of UAV
platforms) and the vibrations involved can comprise a good
calibration of the sensor. Subsequently, on post-processed
data, it is possible to calculate narrowband indices such as
chlorophyll absorption ratio index (CARI), greenness index
(GI), greenness vegetation index (GVI), modified chloro-
phyll absorption ratio index (MCARI), modified normalized
difference vegetation index (MNDVI), simple ratio (SR),
transformed chlorophyll absorption ratio index (TCARI),
triangular vegetation index (TVI), modified vegetation
stress ratio (MVSR), modified soil-adjusted vegetation index
(MSAVI) and photochemical reflectance index (PRI) [94].
Applications ofUAVs toweed management
UAVs are ideal to identify weed patches. e main advan-
tages of UAVs compared to UTVs are the shorter moni-
toring/surveying time they require and optimal control
in the presence of obstacles, which is critical when work-
ing between crop rows [95]. In a few minutes, UAVs can
cover many hectares flying over the field, thus providing
the photographic material for weed patches identification
[61]. ese images are processed via deep neural network
[78], convolutional neural network, and object-based
image analysis [96, 97]. Based on a systematic review of
the literature concerning weed identification by UAVs,
it can be concluded that mainly three types of cameras
are used for weed patches identification: RGB, multispec-
tral and hyperspectral cameras (Table4). ese cameras
Table 3 Hyperspectral sensors andtheir main characteristics
Camera model Lens Spectral range [µm] Spectral bands
[number andµm] Weight [kg] Price (approx.) [€]
CUBERT Snapshot + PA N 450–995 125 (8 µm) ca. 0.5 ca. 50000
Cornirg microHSI 410 SHARK CCD/CMOS 400–1000 300 (2 µm) ca. 0.7 –
Rikola Ltd. hyperspectral camera CMOS 500–900 40 (10 µm) ca. 0.6 ca. 40000
Specim-AISA KESTREL16 Push-broom 600–1640 350 (3 – 8 µm) ca. 2.5 –
Headwall Photonics
Micro-hyperspec X-series NIR InGaAs 900–1700 62 (12.9 µm) ca. 1.1 –
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 6 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
Table 4 Weed patches identication bydierent types ofcamera (multispectral, RGB, hyperspectral)
Crop Weed (common name) Weed (scientic name) Type ofcamera Main results References
Palmer amaranth Amaranthus palmeri Hyperspectral camera Discriminate glyphosate-
resistant from glyphosate-
sensitive weeds
[110]
Spotted knapweed
Babysbreath Centaurea maculosa
Gypsophila paniculata
Hyperspectral camera Detection invasive species
affecting forests, range-
lands, and pastures
[111]
Bunchgrass
Egyptian crowfoot grass
False amaranth
Awnless barnyard grass
Phalaris minor
Dactyloctenium aegyptium
Digera arvensis
Echinochloa colona
RGB camera Identify different weeds [112]
Ragwort Jacobaea vulgaris (Senecio
jacobaea)
Multispectral camera Discriminate weeds in
pastures [113]
Buffel Grass
Spinifex Cenchrus ciliaris
Triodia sp.
RGB camera Discriminate two different
weeds [114]
Beta vulgaris
Zea mays
Hordeum vulgare
Lens esculenta
Pisum sativum
Phaseolus vulgaris
Carthamus tinctorius
Cicer arietinum
Kochia
Marestail
Common lambsquarters
Bassia scoparia
Conyza canadensis
Chenopodium album
Hyperspectral camera Discriminate glyphosate
and dicamba resistant
genotypes from sensitive
genotypes
[115]
Triticum spp.
Triticosecale RGB camera Comparison of cereal geno-
types [116]
Beta vulgaris Weeds Multispectral camera Discriminate crop vs weeds [98]
Beta vulgaris Weeds Multispectral camera Discriminate crop vs weeds [85]
Beta vulgaris Thistle Cirsium arvense Multispectral camera Discriminate crop vs weeds [117]
Beta vulgaris Thistle
Wild buckwheat
Ryegrass
Cirsium arvense
Fallopia convolvulus
Lolium multiflorum
Multispectral camera Discriminate crop vs weeds [101]
Beta vulgaris Thistle Cirsium arvense Multispectral camera Discriminate crop vs weeds [112]
Cicer arietinum Weeds Hyperspectral camera Discriminate crop vs weeds [118]
Glycine max Palmer amaranth
Barnyardgrass
Large crabgrass
Amaranthus palmeri
Echinochloa crus-galli
Digitaria sanguinalis
RGB camera
Multispectral camera Assessment of crop injury
from dicamba [102]
Heliathus annuus Pigweed
Mustard
Bindweed
Lambsquarters
Amaranthus blitoides
Sinapis arvensis
Convolvulus arvensis L
Chenopodium album L
RGB camera
Multispectral camera Discriminate crop vs weeds [64]
Hordeum vulgare Thistle
Coltsfoot Cirsium arvense
Tussilago farfara
RGB camera Discriminate crop vs weeds [119]
Hordeum vulgare Thistle Cirsium arvense RGB camera Discriminate crop vs weeds [99]
Hordeum vulgare Thistle Cirsium arvense RGB camera Discriminate crop vs weeds [100]
Lactuca sativa Common groundsel
Shepherd’s purse
Sow thistle
Senecio vulgaris
Capsella bursa pastoris
Sonchus spp.
Multispectral camera Discriminate crops vs weeds [120]
Sorghum spp. Amaranth
Pigweed
Barnyard grass
Mallow
Nut grass
Fat Hen
Amaranthus macrocarpus
Portulaca oleracea
Echinochloa crus-galli
E. colona
Malva spp.
Cyperus rotundus
Chenopodium album
Hyperspectral camera Discriminate crop vs weeds [121]
Triticum durum Wild oat
Canarygrass
Ryegrass
Avena sterilis
Phalaris canariensis
Lolium rigidum
Multispectral camera Discriminate crop vs weeds [122]
Triticum durum Wild oat
Canarygrass
Ryegrass
Avena fatua
Phalaris canariensis
Lolium rigidum
Hyperspectral camera
Multispectral camera Discriminate crop vs weeds [105]
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 7 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
are very similar in terms of information obtained for the
purpose of weeds identification. Indeed, the three cam-
era types can recognize weed patches with good accuracy
depending on flying altitude, camera resolution and UAV
used. UAVs have been mainly tested on important crops
such as Triticum spp., Hordeum vulgare, Beta vulgaris,
Zea mays [98–101]. ese are among the most cultivated
crops worldwide and are highly susceptible to weed com-
petition especially in early phenological stages. In these
crops, it was possible to identify several dicotyledonous
weeds including Amaranthus palmeri, Chenopodium
album and Cirsium arvense [102–104], as well as differ-
ent monocotyledonous such as Phalaris spp., Avena spp.
and Lolium spp. [105, 106]. ese weed species are wide-
spread globally and can be a serious threat to different
crops [107, 108]. erefore, the combined use of UAVs
and image processing technologies may contribute to
effectively control different weed species interfering with
the crops with relevant environmental benefits [28, 109].
Conclusion
e use of UAVs and machine learning techniques
allow for the identification of weed patches in a cul-
tivated field with accuracy and can improve weed
management sustainability [97]. Weed patches iden-
tification by UAVs can facilitate integrated weed man-
agement (IWM), reducing both the selection pressure
vs herbicide-resistant weeds and herbicides diffusion
in the environment [64]. Recent research has shown
that new technologies are able to discern single weed
species in open fields [63, 106, 126]. If integrated with
weed management planning, this information gathered
via remote imaging analysis can contribute to sustain-
ably improve weed management. In addition, imag-
ing analysis can help in the study of weed dynamics
in the field, as well as their interaction with the crop,
which both represent a necessary step to define new
strategies for weed management based on interspe-
cific crop–weed interactions [127–129]. Recent studies
demonstrate that some weed communities are actually
not detrimental to crop yield and quality [127, 128].
In winter wheat cultivation, a highly diversified weed
community caused lower yield losses than a less diver-
sified one [129]. In soybean, through a combination of
field experiments in which weed species were manipu-
lated in composition and abundance, it has been shown
that increasing levels of weed competition resulted in
an increase in seed protein content without impairing
yield [130].
Most likely, the integration of known and emerging
technologies in this field will greatly improve the sustain-
ability of weed control, following the SSWM approach.
By image analysis, different machine learning techniques
will be able to provide a reliable overview of the level and
type of infestation. Specific algorithms can be trained to
manage weeds removal by Autonomous Weeding Robot
(AWR), via herbicide spray or mechanical means [131].
Also, the creation of a specific weed images dataset is
crucial to achieve this goal. is approach must neces-
sarily rely on a dataset of photographs taken in dedicated
experimental fields, labeled in extended COCO/POCO
(Common Objects in COntext/ Plant Objects in COn-
text) format [86] and integrated with images from Plant-
Village dataset [83] or other existing ones.
New insights on weed population dynamics and their
competition with crops are needed in order to extend
this approach to real agricultural contexts, so as to specif-
ically recognize and eliminate only harmful weed species.
e overall objective is to overcome the consequences
of biological vacuum around the crop, which has been
proved to be highly impacting for both biotic and the
abiotic components of the environment [132, 133], with
long-term consequences on human safety on earth.
These types of cameras can be assembled with drones for reaching SSWM purposes
Table 4 (continued)
Crop Weed (common name) Weed (scientic name) Type ofcamera Main results References
Triticum sp. Thistle Cirsium arvense RGB camera Discriminate crop vs weeds [99]
Triticum spp. Weeds Hyperspectral camera Discriminate crop vs weeds [118]
Vitis vinifera Bermuda grass Cynodon dactylon RGB camera Discriminate crop vs weeds [123]
Zea mays Weeds Multispectral camera Discriminate crop vs weeds [124]
Zea mays Common lambsquarters
Thistle Chenopodium album
Cirsium arvense
Multispectral camera Discriminate monocotyle-
dons (crops) vs dicotyle-
dons (weeds)
[104]
Zea mays Common lambsquarters
Thistle Chenopodium album
Cirsium arvense
Multispectral camera Discriminate crop vs weeds [101]
Zea mays Mat amaranth
Johnsongrass Amaranthus blitoides
Sorghum halepense
Multispectral camera Discriminate crop vs weeds [125]
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 8 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
Abbreviations
IWM: Integrated Weed Management; UAV: Unmanned Aerial Vehicles; UTV:
Unmanned Terrestrial Vehicles; SSWM: Specific Weed Management; RGB: Red
Green Blue; VIS: Visible; GRVI: Green/Red Vegetation Index; GI: Greenness Index;
ExG: Excessive Greenness; GPS: Global Positioning System; CARI: Chlorophyll
Absorption Ratio Index; MCARI: Modified Chlorophyll Absorption Ratio
Index; MNDVI: Modified Normalized Difference Vegetation Index; SR: Simple
Ratio; TCARI: Transformed Chlorophyll Absorption Ratio Index; TVI: Triangular
Vegetation Index; MVSR: Modified Vegetation Stress Ratio; MSAVI: Modified
Soil-Adjusted Vegetation Index; PRI: Photochemical Reflectance Index; AWR
: Autonomous Weeding Robot; COCO/POCO: Common Objects in COntext/
Plant Objects in COntext.
Acknowledgements
Not applicable.
Authors’ contributions
ME and MC wrote the original draft and produced all the tables and the
figures. AM, FS and VC revised the text for the final version. VC and ME con-
ceived the idea of writing the review. All authors read and approved the final
manuscript.
Funding
This research was funded by MEDES Foundation.
Availability of data and materials
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Received: 7 January 2021 Accepted: 15 February 2021
References
1. Oerke EC. Crop losses to pests. Sci: J Agric; 2006.
2. Guglielmini AC, Verdú AMC, Satorre EH. Competitive ability of five
common weed species in competition with soybean. Int J Pest Manag.
2017;63:30–6.
3. Kaur H, Brar SG, Shete PP. A REVIEW on different weed management
approaches. Int J Curr Microbiol Appl Sci. 2019; 8: 2854.
4. Korav S, Ram V, Ray LIP, Krishnappa R, Singh NJ, Premaradhya N. Weed
pressure on growth and yield of groundnut (Arachis hypogaea L.) in
Meghalaya, India. Int J Curr Microbiol Appl Sci. 2018;7:2852–8.
5. Thompson CR, Dille JA, Peterson DE. Weed Competition and Manage-
ment in Sorghum. In Sorghum. Ciampitti IA, Vara Prasad P, eds. 2019.
6. Zohaib A, Abbas T, Tabassum T. Weeds cause losses in field crops
through allelopathy. Not Sci Biol. 2016;8:47–56.
7. Cirillo V, Masin R, Maggio A, Zanin G. Crop-weed interactions in saline
environments. Eur J Agron. 2018;99:51–61.
8. Hauvermale AL, Sanad MNME. Phenological plasticity of wild and
cultivated plants. Plant communities and their environment. London:
IntechOpen; 2019.
9. Guo L, Qiu J, Li L-F, Lu B, Olsen K, Fan L. Genomic clues for crop-weed
interactions and evolution. Trends Plant Sci. 2018;23:1102–15.
10. Kumar A, Choudhary T, Das S, Meena SK. Weed Seed Bank: impacts and
management for future crop production BT—agronomic crops, Vol 2.
Management Practices. Hasanuzzaman M, editor. 2019. p. 207–23.
11. Smith JD, Dubois T, Mallogo R, Njau E-F, Tua S, Srinivasan R. Host range
of the invasive tomato pest tuta absoluta meyrick (Lepidoptera:
Gelechiidae) on solanaceous crops and weeds in Tanzania. Fla Entomol.
2019;101:573.
12. Roshan P, Kulshreshtha A, Hallan V. Global weed-infecting geminiviruses
BT-geminiviruses: impact, challenges and approaches. Berlin: Springer;
2019. p. 103–21.
13. Srinivasan R, Cervantes FA, Alvarez JM. Aphid-borne virus dynamics in
the potato-weed pathosystem. Insect Pests Potato. 2013.
14. Eshed N, Wahl I. Role of wild grasses in epidemics of powdery mildew.
Phytopathology. 1975;65:57–63.
15. Adkins S, Shabbir A. Biology, ecology and management of the invasive
parthenium weed (Parthenium hysterophorus L.). Pest Manag Sci.
2014;70:1023–9.
16. Tanveer A, Khaliq A. Mahajan G. Interference and management of par-
thenium: The world ’ s most important invasive weed. J Crop Prot; 2015.
p. 68.
17. Khan MR, Somvanshi VS, Rao U. Emerging nematode pest of rice, wheat
and onion. Rice Root-Knot Nematode Popular Kheti. 2017;3:53–5.
18. Ersin A, Gargin S, Esitken A, Guzel NP, Atay AN, Altindal M, et al. The
effect of weed competition on apple fruit quality. Not Bot Horti
Agrobot Cluj-Napoca. 2017;45:120–5.
19. Yadav T, Chopra NK, Chopra NK, Kumar R, Soni PG. Assessment of
critical period of crop-weed competition in forage cowpea (Vigna
unguiculata) and its effect on seed yield and quality. Indian J Agron.
2018;63:124–7.
20. Hall CW. Vacuum treatment of milk. in: vacuum technology transac-
tions. Pergamon; 2013.
21. Sabra FS, Mahmoud MS. Utilization of herbicidal treatments to
overcome weed problems in utilization of herbicidal treatments to
overcome weed problems in peppermint (Mentha piperita L.) cultiva-
tion under Egyptian conditions. Asian J Agric Food Sci. 2015;03:1–7.
22. Mantle P. Comparative ergot alkaloid elaboration by selected plecten-
chymatic mycelia of Claviceps purpurea through sequential cycles of
axenic culture and plant parasitism. Biology. 2020;9:41.
23. Kumar R, Katiyar R, Kumar S, Kumar T, Singh V. Lantana camara: an alien
weed, its impact on animal health and strategies to control. J Exp Biol
Agric Sci. 2016;4:321–37.
24. Denisow-Pietrzyk M, Pietrzyk Ł, Denisow B. Asteraceae species as
potential environmental factors of allergy. Environ Sci Pollut Res.
2019;26:6290–300.
25. Allan S, Shi B, Adkins SW. Impact of parthenium weed on human and
animal health. Parthenium Weed Biol Ecol Manag. 2018;7:105.
26. Korres NE, Burgos NR, Travlos I, Vurro M, Gitsopoulos TK, Varanasi
VK, et al. New directions for integrated weed management: mod-
ern technologies, tools and knowledge discovery. Adv Agron.
2019;155:243–319.
27. Partel V, Charan Kakarla S, Ampatzidis Y. Development and evaluation
of a low-cost and smart technology for precision weed management
utilizing artificial intelligence. Comput Electron Agric. 2019;157:339–50.
28. Hunter JE, Gannon TW, Richardson RJ, Yelverton FH, Leon RG.
Integration of remote-weed mapping and an autonomous spraying
unmanned aerial vehicle for site-specific weed management. Pest
Manag Sci. 2020;76:1386–92.
29. Soltani N, Dille JA, Burke IC, Everman WJ, VanGessel MJ, Davis VM, et al.
Potential Corn yield losses from weeds in North America. Weed Technol.
2016;30:979–84.
30. Llewellyn R, Ronning D, Ouzman J, Walker S, Mayfield A, Clarke M.
Impact of Weeds on Australian Grain Production: the cost of weeds to
Australian grain growers and the adoption of weed management and
tillage practices Report for GRDC. CSIRO. 2016.
31. Appleby AP, Müller F, Carpy S. Weed control. In Ullmann’s Encyclope-
dia of Industrial Chemistry, (Ed.). https ://doi.org/10.1002/14356 007.
a28_165
32. Swanton CJ, Nkoa R, Blackshaw RE. Experimental methods for crop-
weed competition studies. Weed Sci. 2015;63:2–11.
33. EUSROSTAT. Statistics Eurostat. https ://ec.europ a.eu/euros tat/datab
rowse r/view/aei_fm_salpe st09/defau lt/table ?lang=en.
34. Alexandratos N, Bruinsma J. World agriculture towards 2030/2050: the
2012 revision. 2012.
35. Ray DK, Mueller ND, West PC, Foley JA. Yield trends are insufficient to
double global crop production by 2050. PLoS ONE. 2013;8:e66428.
36. Mall RK, Gupta A, Sonkar G. 2—effect of climate change on agricultural
crops. In: Dubey SK, Pandey A, Sangwan RSBT-CD in B and B, editors.
Elsevier. 2017. p. 23–46.
37. Westwood JH, Charudattan R, Duke SO, Fennimore SA, Marrone P,
Slaughter DC, et al. Weed management in 2050: perspectives on the
future of weed science. Weed Sci. 2018;66:275–85.
38. Dong W, Liu E, Yan C, Tian J, Zhang H, Zhang Y. Impact of no tillage
vs. conventional tillage on the soil bacterial community structure in
a winter wheat cropping succession in northern China. Eur J Soil Biol.
2017;80:35–42.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 9 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
39. Duke S, Heap I. Evolution of weed resistance to herbicides: what have
we learned after 70 years. Biol Physiol Mol Biol Weeds. 2017;63–86.
40. Hicks HL, Comont D, Coutts SR, Crook L, Hull R, Norris K, et al. The fac-
tors driving evolved herbicide resistance at a national scale. Nat Ecol
Evol. 2018;2:529–36.
41. Büchi L, Wendling M, Amossé C, Necpalova M, Charles R. Importance
of cover crops in alleviating negative effects of reduced soil tillage and
promoting soil fertility in a winter wheat cropping system. Agric Ecosyst
Environ. 2018;256:92–104.
42. Jackson L, Calderon F, Steenwerth K, Scow K, Rolston D. Responses of
soil microbial processes and community structure to tillage events and
implications for soil quality. Geoderma. 2003;114:305–17.
43. Muola A, Fuchs B, Laihonen M, Rainio K, Heikkonen L, Ruuskanen S,
et al. Risk in the circular food economy: glyphosate-based herbicide
residues in manure fertilizers decrease crop yield. Sci Total Environ.
2020;750:141422.
44. Riemens MM, Van Der Weide RY, Bleeker PO, Lotz LAP. Effect of stale
seedbed preparations and subsequent weed control in lettuce (cv.
Iceboll) on weed densities. Weed Res. 2007;47:149–56.
45. Bajwa AA, Walsh M, Chauhan BS. Weed management using crop com-
petition in Australia. Crop Prot. 2017;95:8–13.
46. Swanton CJ, Weise SF. Integrated weed management: the rationale and
approach. Weed Technol. 1991;5:657–63.
47. Young SL, Pitla SK, Van Evert FK, Schueller JK, Pierce FJ. Moving inte-
grated weed management from low level to a truly integrated and
highly specific weed management system using advanced technolo-
gies. Weed Res. 2017;57:1–5.
48. Bonny S. Genetically modified herbicide-tolerant crops, weeds, and
herbicides: overview and impact. Environ Manage. 2016;57:31–48.
49. Rose MT, Cavagnaro TR, Scanlan CA, Rose TJ, Vancov T, Kimber S, et al.
Impact of Herbicides on Soil Biology and Function. In: Sparks DL, editor.
Adv Agro. Academic Press. 2016; pp. 133–220.
50. Shannon D, Clay DE, Sudduth KA. An introduction to precision agricul-
ture. Precis Agric Basics. 2018.
51. Mahlein AK. Plant disease detection by imaging sensors—parallels
and specific demands for precision agriculture and plant phenotyping.
Plant Dis. 2016;100:241–51.
52. Nawar S, Corstanje R, Halcro G, Mulla D, Mouazen AM. Delineation of
soil management zones for variable-rate fertilization: a review. Adv
Agron. 2017;143:175–245.
53. Lama GFC, Errico A, Francalanci S, Solari L, Preti F, Chirico GB. Evalua-
tion of flow resistance models based on field experiments in a partly
vegetated reclamation channel. Geosciences. 2020;10:47.
54. Gautam PV, Kushwaha HL, Kumar A, Kumar D. Mechatronics appli-
cation in precision sowing: a review. Int J Curr Microbiol Appl Sci.
2019;8:1793–807.
55. Chlingaryan A, Sukkarieh S, Whelan B. Machine learning approaches
for crop yield prediction and nitrogen status estimation in precision
agriculture: a review. Comput Electronics Agric. 2018;151:61–9.
56. Murray SC. Optical sensors advancing precision in agricultural produc-
tion. Photonics Spectra. 2017.
57. Dyshekov AI, Smirnov IG, Mirzaev MA, Shereuzhev MA. Principles of
functioning of the autonomous device for weed control for precision
agriculture. IOP Conf Ser Mater Sci Eng. 2020.
58. Duhan JS, Kumar R, Kumar N, Kaur P, Nehra K, Duhan S. Nanotechnol-
ogy: the new perspective in precision agriculture. Biotechnol Rep.
2017;15:11–23.
59. Mogili UR, Deepak BBVL. Review on application of drone systems in
precision agriculture. Procedia Comput Sci. 2018;133:502–9.
60. Raj R, Kar S, Nandan R, Jagarlapudi A. precision agriculture and
unmanned aerial vehicles (UAVs) BT unmanned aerial vehicle: applica-
tions in agriculture and environment. In: Avtar R, Watanabe T, editors.
Cham: Springer International Publishing; 2020.
61. Krishna KR. Agricultural drones: a peaceful pursuit. Taylor Francis. 2018.
62. John KN, Valentin V, Abdullah B, Bayat M, Kargar MH, Zargar M. Weed
mapping technologies in discerning and managing weed infestation
levels of farming systems. Res Crops. 2020.
63. Lottes P, Khanna R, Pfeifer J, Siegwart R, Stachniss C. UAV-based crop
and weed classification for smart farming. Proc IEEE Int Conf Robot
Autom. 2017.
64. López-Granados F, Torres-Sánchez J, Serrano-Pérez A, de Castro AI,
Mesas-Carrascosa Fco-J, Peña J-M. Early season weed mapping in sun-
flower using UAV technology: variability of herbicide treatment maps
against weed thresholds. Precis Agric. 2016;17:183–99.
65. Jiménez-Brenes FM, López-Granados F, Torres-Sánchez J, Peña JM,
Ramírez P, Castillejo-González IL, et al. Automatic UAV-based detection
of Cynodon dactylon for site-specific vineyard management. PLoS ONE.
2019.
66. Daponte P, Vito LD, Glielmo L, Iannelli L, Liuzza D, Picariello F, et al. A
review on the use of drones for precision agriculture. IOP Conf Ser Earth
Environ Sci. 2019;275:012022.
67. Radoglou-Grammatikis P, Sarigiannidis P, Lagkas T, Moscholios I. A com-
pilation of UAV applications for precision agriculture. Comput Netw.
2020;172:107148.
68. Mancini A, Frontoni E, Zingaretti P. Satellite and UAV data for Preci-
sion Agriculture Applications. In: 2019 International Conference on
Unmanned Aircraft Systems (ICUAS). IEEE. 2019.
69. Brook A, De Micco V, Battipaglia G, Erbaggio A, Ludeno G, Catapano I,
et al. A smart multiple spatial and temporal resolution system to sup-
port precision agriculture from satellite images: proof of concept on
Aglianico vineyard. Remote Sens Environ. 2020;240:111679.
70. Watts AC, Ambrosia VG, Hinkley EA. Unmanned aircraft systems in
remote sensing and scientific research: classification and considerations
of use. Remote Sens. 2012;4:1671–92.
71. Manfreda S, McCabe ME, Miller PE, Lucas R, Madrigal VP, Mallinis G, et al.
On the use of unmanned aerial systems for environmental monitoring.
Remote Sens. 2018;10:641.
72. Anderson K, Gaston KJ. Lightweight unmanned aerial vehicles will
revolutionize spatial ecology. Front Ecol Environ. 2013;11:138–46.
73. Whitehead K, Hugenholtz CH. Remote sensing of the environment with
small unmanned aircraft systems (UASs), part 1: a review of progress
and challenges. J Unmanned Veh Syst. 2014;02:69–85.
74. Torres-Sánchez J, Peña JM, de Castro AI, López-Granados F. Multi-tem-
poral mapping of the vegetation fraction in early-season wheat fields
using images from UAV. Comput Electron Agric. 2014;103:104–13.
75. Jannoura R, Brinkmann K, Uteau D, Bruns C, Joergensen RG. Monitor-
ing of crop biomass using true colour aerial photographs taken from a
remote controlled hexacopter. Biosyst Eng. 2015;129:341–51.
76. Motohka T, Nasahara KN, Oguma H, Tsuchida S. Applicability of green-
red vegetation index for remote sensing of vegetation phenology.
Remote Sens. 2010;2:2369–87.
77. Xue J, Su B. Significant remote sensing vegetation indices: a review of
developments and applications. J Sens. 2017.
78. Crimaldi M, Cristiano V, De Vivo A, Isernia M, Ivanov P, Sarghini F. Neural
network algorithms for real time plant diseases detection using UAVs.
In: Coppola A, Di Renzo GC, Altieri G, D’Antonio P, editors. Innov Biosyst
Eng Sustain Agric For Food Prod. 2020. p. 827–35.
79. Pflanz M, Schirrmann M, Nordmeyer H. Drone based weed monitoring
with an image feature classifier. Julius-Kühn-Archiv. 2018;458:379–84.
80. Etienne A, Saraswat D. Machine learning approaches to automate
weed detection by UAV based sensors. In: Autonomous air and ground
sensing systems for agricultural optimization and phenotyping IV.
International Society for Optics and Photonics, 2019. p. 110080R.
81. Wang S, Han Y, Chen J, Pan Y, Cao Y, Meng H, et al. A deep-learning-
based low-altitude remote sensing algorithm for weed classification
in ecological irrigation area. In: International Conference on Intelligent
Technologies and Applications. Springer, Singapore, 2018. p. 451–460.
82. Bah MD, Hafiane A, Canals R. Deep learning with unsupervised data
labeling for weed detection in line crops in UAV images. Remote Sens.
2018;10:1690.
83. Hughes DP, Salathe M. An open access repository of images on plant
health to enable the development of mobile disease diagnostics.
ArXiv151108060 Cs. 2016. http://arxiv .org/abs/1511.08060
84. Singh D, Jain N, Jain P, Kayal P, Kumawat S, Batra N. PlantDoc: a dataset
for visual plant disease detection. In: Proceedings of the 7th ACM IKDD
CoDS and 25th COMAD. 2020. p. 249–253.
85. Sa I, Popović M, Khanna R, Chen Z, Lottes P, Liebisch F, et al. WeedMap: a
large-scale semantic weed mapping framework using aerial multispec-
tral imaging and deep neural network for precision farming. Remote
Sens. 2018;10:1–25.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 10 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
86. Wspanialy P, Brooks J, Moussa M. An image labeling tool and agricul-
tural dataset for deep learning. ArXiv200403351 Cs Eess. 2020. http://
arxiv .org/abs/2004.03351
87. Micasense Inc. Automatic Calibration Panel Detection (QR Mode).
MicaSense Knowl. Base. 2020. https ://suppo rt.micas ense.com/hc/
en-us/artic les/36001 86187 74
88. Micasense Inc. Best practices: Collecting Data with MicaSense Sensors.
MicaSense Knowl. Base. 2020. https ://suppo rt.micas ense.com/hc/
en-us/artic les/22489 3167.
89. Laliberte AS, Goforth MA, Steele CM, Rango A. Multispectral
remote sensing from unmanned aircraft: image processing work-
flows and applications for rangeland environments. Remote Sens.
2011;3:2529–51.
90. Barrero O, Perdomo SA. RGB and multispectral UAV image fusion for
Gramineae weed detection in rice fields. Agric: Precis; 2018.
91. Candiago S, Remondino F, De Giglio M, Dubbini M, Gattelli M. Evaluat-
ing multispectral images and vegetation indices for precision farming
applications from UAV images. Remote Sens. 2015;7:4026.
92. Mazzia V, Comba L, Khaliq A, Chiaberge M, Gay P. UAV and machine
learning based refinement of a satellite-driven vegetation index for
precision agriculture. Sensors. 2020;20:2530.
93. Brook A, Ben-Dor E. Supervised Vicarious Calibration (SVC) of
multi-source hyperspectral remote-sensing data. Remote Sens.
2015;7:6196–223.
94. Adao T, Hruska J, Padua L, Bessa J, Peres E, Morais R, et al. Hyperspectral
imaging: a review on UAV-based sensors, data processing and applica-
tions for agriculture and forestry. Remote Sens. 2017;9:1110.
95. Hassanein M, El-Sheimy N. An efficient weed detection procedure using
low-cost UAV imagery system for precision agriculture applications. Int
Arch Photogramm Remote Sens Spat Inf Sci ISPRS Arch. 2018;42:181–7.
96. Tsouros DC, Bibi S, Sarigiannidis PG. A review on UAV-based applica-
tions for precision agriculture. Inf Switz. 2019;10:349.
97. Maes WH, Steppe K. Perspectives for remote sensing with unmanned
aerial vehicles in precision agriculture. Trends Plant Sci. 2019;24:152–64.
98. Fawakherji M, Potena C, Bloisi DD, Imperoli M, Pretto A, Nardi D. UAV
Image based crop and weed distribution estimation on embedded
GPU boards BT computer analysis of images and patterns. Berlin:
Springer; 2019.
99. Rasmussen J, Nielsen J, Streibig JC, Jensen JE, Pedersen KS, Olsen SI.
Pre-harvest weed mapping of Cirsium arvense in wheat and barley with
off-the-shelf UAVs. Precis Agric. 2019;20:983–99.
100. Franco C, Guada C, Rodríguez JT, Nielsen J, Rasmussen J, Gómez D, et al.
Automatic detection of thistle-weeds in cereal crops from aerial RGB
images BT information processing and management of uncertainty in
knowledge-based systems. Applications. 2018. p. 441–52.
101. Louargant M, Jones G, Faroux R, Paoli JN, Maillot T, Gée C, et al.
Unsupervised classification algorithm for early weed detection in
row-crops by combining spatial and spectral information. Remote Sens.
2018;10:1–18.
102. Huang Y, Reddy KN, Fletcher RS, Pennington D. UAV low-altitude remote
sensing for precision weed management. Weed Technol. 2018;32:2–6.
103. Hansen KD, Garcia-Ruiz F, Kazmi W, Bisgaard M, La Cour-Harbo A, Ras-
mussen J, et al. An autonomous robotic system for mapping weeds in
fields. IFAC Proc Vol IFAC-Pap. 2013;8:217–24.
104. Louargant M, Villette S, Jones G, Vigneau N, Paoli JN, Gée C. Weed
detection by UAV: simulation of the impact of spectral mixing in multi-
spectral images. Precis Agric. 2017;1–20.
105. López-Granados F, Jurado-Expósito M, Peña-Barragán JM, García-Torres
L. Using remote sensing for identification of late-season grass weed
patches in wheat. Weed Sci. 2006;54:346–53.
106. Sanders JT, Everman WJ, Austin R, Roberson GT, Richardson RJ. Weed
species differentiation using spectral reflectance land image classifica-
tion. ProcSPIE. 2019.
107. Davis S, Mangold J, Menalled F, Orloff N, Miller Z, Lehnhoff E. A Meta-
analysis of Canada Thistle (Cirsium arvense) management. Weed Sci.
2018;66:548–57.
108. Alshallash KS. Germination of weed species (Avena fatua, Bromus
catharticus, Chenopodium album and Phalaris minor) with implications
for their dispersal and control. Ann Agric Sci. 2018;63:91–7.
109. De Castro AI, Torres-Sánchez J, Peña JM, Jiménez-Brenes FM, Csillik
O, López-Granados F. An automatic random forest-OBIA algorithm
for early weed mapping between and within crop rows using UAV
imagery. Remote Sens. 2018;10:285.
110. Reddy KN, Huang Y, Lee MA, Nandula VK, Fletcher RS, Thomson SJ, et al.
Glyphosate-resistant and glyphosate-susceptible Palmer amaranth
(Amaranthus palmeri S. Wats.): hyperspectral reflectance properties of
plants and potential for classification. Pest Manag Sci. 2014;70:1910–7.
111. Lass LW, Prather TS, Glenn NF, Weber KT, Mundt JT, Pettingill J. A
review of remote sensing of invasive weeds and example of the early
detection of spotted knapweed (Centaurea maculosa) and babys-
breath (Gypsophila paniculata) with a hyperspectral sensor. Weed Sci.
2005;53:242–51.
112. Tiwari O, Goyal V, Kumar P, Vij S. An experimental set up for utilizing
convolutional neural network in automated weed detection. Proc 2019
4th Int Conf Internet Things Smart Innov Usages IoT-SIU 2019. 2019;1–6.
113. Tay JYL, Erfmeier A, Kalwij JM. Reaching new heights: can drones
replace current methods to study plant population dynamics? Plant
Ecol. 2018;219:1139–50.
114. Sandino J, Gonzalez F. A Novel Approach for invasive weeds and veg-
etation surveys using UAS and artificial intelligence. 2018 23rd Int Conf
Methods Models Autom Robo. 2018.
115. Scherrer B, Sheppard J, Jha P, Shaw JA. Hyperspectral imaging and
neural networks to classify herbicide-resistant weeds. J Appl Remote
Sens. 2019;13:1.
116. Svensgaard J, Jensen SM, Westergaard JC, Nielsen J, Christensen S,
Rasmussen J. Can reproducible comparisons of cereal genotypes be
generated in field experiments based on UAV imagery using RGB
cameras? Eur J Agron. 2019;106:49–57.
117. Garcia-Ruiz FJ, Wulfsohn D, Rasmussen J. Sugar beet (Beta vulgaris L.)
and thistle (Cirsium arvensis L.) discrimination based on field spectral
data. Biosyst Eng. 2015;139:1–15.
118. Shapira U, Herrmann I, Karnieli A, Bonfil JD. Weeds detection by ground-
level hyperspectral data. ISPRS Archive Vol. XXXVIII. 2010;38:27–33.
119. Rasmussen J, Nielsen J, Garcia-Ruiz F, Christensen S, Streibig JC. Poten-
tial uses of small unmanned aircraft systems (UAS) in weed research.
Weed Res. 2013;53:242–8.
120. Slaughter DC, Giles DK, Fennimore SA, Smith RF. Multispectral machine
vision identification of lettuce and weed seedlings for automated weed
control. Weed Technol. 2008;22:378–84.
121. Norasma N, Ya C, George D. Spectral discrimination of weeds using
hyperspectral. 2013.
122. López-Granados F, Peña-Barragán JM, Jurado-Expósito M, Francisco-
Fernández M, Cao R, Alonso-Betanzos A, et al. Multispectral classifi-
cation of grass weeds and wheat (Triticum durum) using linear and
nonparametric functional discriminant analysis and neural networks.
Weed Res. 2008;48:28–37.
123. de Castro AI, Peña JM, Torres-Sánchez J, Jiménez-Brenes F, López-Gra-
nados F. Mapping Cynodon dactylon in vineyards using UAV images for
site-specific weed control. Adv Anim Biosci. 2017;8:267–71.
124. Fawakherji M, Potena C, Bloisi DD, Imperoli M, Pretto A, Nardi D. UAV
image based crop and weed distribution estimation on embedded
GPU boards. Commun Comput Inf Sci. 2019;1089:100–8.
125. Peña JM, Torres-Sánchez J, de Castro AI, Kelly M, López-Granados F.
Weed mapping in early-season maize fields using object-based analysis
of unmanned aerial vehicle (UAV) images. PLoS ONE. 2013;8:1–11.
126. Pantazi X-E, Moshou D, Bravo C. Active learning system for weed
species recognition based on hyperspectral sensing. Biosyst Eng.
2016;146:193–202.
127. Gibson DJ, Young BG, Wood AJ. Can weeds enhance profitability?
Integrating ecological concepts to address crop-weed competition and
yield quality. J Ecol. 2017;105:900–4.
128. Millar KDL, Ebbs SD, Gibson DJ, Wood AJ, Young BG. Evaluation of
physiological parameters for the prediction of seed yield and quality for
soybean (Glycine max) plants grown in the presence of weed competi-
tion. Plant Biosyst. 2011;145:1–11.
129. Adeux G, Vieren E, Carlesi S, Bàrberi P, Munier-Jolain N, Cordeau S.
Mitigating crop yield losses through weed diversity. Nat Sustain.
2019;2:1018–26.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Page 11 of 11
Espositoetal. Chem. Biol. Technol. Agric. (2021) 8:18
130. Millar K, Gibson DJ, Young BG, Wood AJ. Impact of interspecific compe-
tition on seed development and quality of five soybean cultivars. Aust J
Exp Agric. 2007;47:1455–9.
131. Talaviya T, Shah D, Patel N, Yagnik H, Shah M. Implementation of artificial
intelligence in agriculture for optimisation of irrigation and application
of pesticides and herbicides. Artif Intell Agric. 2020;4:58–73.
132. Lanz B, Dietz S, Swanson T. The expansion of modern agriculture and
global biodiversity decline: an integrated assessment. Ecol Econ.
2018;144:260–77.
133. Erisman JW, van Eekeren N, de Wit J, Koopmans C, Cuijpers W, Oerle-
mans N, et al. Agriculture and biodiversity: a better balance benefits
both. AIMS Agric Food. 2016;1:157–74.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in pub-
lished maps and institutional affiliations.
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
1.
2.
3.
4.
5.
6.
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
accessing, sharing, receiving or otherwise using the Springer Nature journal content you agree to these terms of use (“Terms”). For these
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
apply.
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
detailed in the Privacy Policy.
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
not:
use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
control;
use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
otherwise unlawful;
falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
writing;
use bots or other automated methods to access the content or redirect messages
override any security feature or exclusionary protocol; or
share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
content.
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
These terms of use are reviewed regularly and may be amended at any time. Springer Nature is not obligated to publish any information or
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at
onlineservice@springernature.com