Hostname: page-component-89b8bd64d-r6c6k Total loading time: 0 Render date: 2026-05-07T18:16:24.015Z Has data issue: false hasContentIssue false

Weed detection to weed recognition: reviewing 50 years of research to identify constraints and opportunities for large-scale cropping systems

Published online by Cambridge University Press:  02 November 2022

Guy R.Y. Coleman*
Affiliation:
Graduate Student, School of Life and Environmental Sciences, The University of Sydney, Brownlow Hill, NSW, Australia
Asher Bender
Affiliation:
Postdoctoral Research Associate, Australian Centre for Field Robotics, The University of Sydney, Chippendale, NSW, Australia
Kun Hu
Affiliation:
Postdoctoral Research Associate, School of Computer Science, The University of Sydney, Sydney, NSW, Australia
Shaun M. Sharpe
Affiliation:
Research Scientist, Agriculture and Agri-Food Canada, Saskatoon, Saskatchewan, Canada
Arnold W. Schumann
Affiliation:
Professor, University of Florida, Citrus Research and Education Center, Lake Alfred, FL, USA
Zhiyong Wang
Affiliation:
Associate Professor, School of Computer Science, The University of Sydney, Sydney, NSW, Australia
Muthukumar V. Bagavathiannan
Affiliation:
Professor, Department of Soil and Crop Sciences, Texas A&M University, College Station, TX, USA
Nathan S. Boyd
Affiliation:
Associate Professor, University of Florida, Gulf Coast Research and Education Center, Wimauma, FL, USA
Michael J. Walsh
Affiliation:
Associate Professor, School of Life and Environmental Sciences, The University of Sydney, Brownlow Hill, NSW, Australia
*
Author for correspondence: Guy Coleman, Graduate Student, The University of Sydney, School of Life and Environmental Sciences, 380 Werombi Road, Brownlow Hill, NSW, 2570 Email: guy.coleman@sydney.edu.au
Rights & Permissions [Opens in a new window]

Abstract

The past 50 yr of advances in weed recognition technologies have poised site-specific weed control (SSWC) on the cusp of requisite performance for large-scale production systems. The technology offers improved management of diverse weed morphology over highly variable background environments. SSWC enables the use of nonselective weed control options, such as lasers and electrical weeding, as feasible in-crop selective alternatives to herbicides by targeting individual weeds. This review looks at the progress made over this half-century of research and its implications for future weed recognition and control efforts; summarizing advances in computer vision techniques and the most recent deep convolutional neural network (CNN) approaches to weed recognition. The first use of CNNs for plant identification in 2015 began an era of rapid improvement in algorithm performance on larger and more diverse datasets. These performance gains and subsequent research have shown that the variability of large-scale cropping systems is best managed by deep learning for in-crop weed recognition. The benefits of deep learning and improved accessibility to open-source software and hardware tools has been evident in the adoption of these tools by weed researchers and the increased popularity of CNN-based weed recognition research. The field of machine learning holds substantial promise for weed control, especially the implementation of truly integrated weed management strategies. Whereas previous approaches sought to reduce environmental variability or manage it with advanced algorithms, research in deep learning architectures suggests that large-scale, multi-modal approaches are the future for weed recognition.

Information

Type
Review
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2022. Published by Cambridge University Press on behalf of the Weed Science Society of America
Figure 0

Table 1. Key definitions for varying levels of specificity in the location and characterization of weeds and the technologies enabling the research and development of site-specific weed control tools.

Figure 1

Figure 1. Publication counts (including journal articles, conference papers, and books) by year for the search term on Scopus: “weed detection” OR “weed recognition” OR “weed identification” indicating the recent rise in popularity. A total of 781 documents were returned beginning in 1989 and ending in 2021; 2022 has been excluded. Columns are colored by corresponding section in this review.

Figure 2

Figure 2. An example image analysis flow for conventional weed detection algorithms to extract ginger plants from the background and to then identify purple nutsedge. The original image is first transformed into the hue, saturation, and value color space, before image features such as mean color channel statistics are calculated, thresholds applied through a deterministic algorithm, resulting in the identification of ginger plants.

Figure 3

Figure 3. The improvement in the performance of plant classification accuracy over time from the first attempts in the mid-1980s through 2022 (n = 67). Data are provided in Supplementary Table S1. Each point represents the top-performing classification accuracy for the top-performing algorithm in the cited article. Where multiple datasets were used to train distinct algorithms, performance was reported separately. Algorithms that relied on conventional (i.e. manual) feature extraction are shown in yellow circles, whereas the automatic, convolutional neural network (CNN)–based feature extraction is indicated by purple circles. Circle size indicates dataset size as a measure of dataset diversity. Results were not included, if papers did not report accuracy. This underrepresents more recent results, which more frequently report metrics such as F1-score, mAP, precision, and recall.

Figure 4

Figure 4. A graphical representation of an artificial neural network (ANN) architecture tested in Burks et al. (2000b) with 11 input features, two hidden layers with 12 and 6 nodes, respectively, and a 6-node output layer. The 11 inputs represent textural features extracted using a color co-occurrence matrix for hue (H) and saturation (S). The values following the letter indicate the texture statistic used: (1) 2nd moment, (2) mean intensity, (4) correlation, (5) product moment, (6) inverse difference, (9) difference entropy, (10) information correlation measure 1, and (11) information correlation measure 2. Example weightings between nodes are represented by color (red: negative; blue: positive) with the intensity of each color indicating the weighting of the connection. Each node (circle) performs a calculation on the incoming information, passing on the outcome to subsequent layers of the network.

Figure 5

Figure 5. In general, there are four possible levels of weed detection and identification based on the implementation of different algorithm architectures: (A) image classification (whole-image level); (B) object detection (localization within an image); (C) semantic segmentation (pixel-wise classification); (D) instance segmentation (pixel and object classification). The development and usage of each is dictated by the desired level of accuracy and application precision, with each method providing a theoretically greater level of information on weed location than the previous.

Figure 6

Table 2. Some of the recent commercial ventures into weed identification for large-scale cropping systems.a

Figure 7

Figure 6. Comparison of model parameter counts and performance on the CottonWeeds dataset. Model architectures perform differently on different image datasets, and model size is not a consistent indicator of likely performance. Chart created with data from Chen et al. (2021).

Figure 8

Table 3. Publicly available and open-access weed image datasets in different crop production systems with URLs for access. Datasets listed as available in published research but without an accessible URL have been excluded. The list expands on datasets provided in Lu and Young (2020) and Hu et al. (2021c). Where segmentation datasets are provided, the total class number does not include an assumed background or soil class. Segmentation includes masked datasets, which can be converted into either semantic or instance segmentation if needed. The Weed-AI platform hosts numerous datasets and serves as an open-source, open-access platform rather than a dataset itself. The Eden Library (Mylonas et al. 2022) is another weed image dataset platform; however, datasets are not open access, and as a result it has not been included in this list.

Supplementary material: File

Coleman et al. supplementary material

Coleman et al. supplementary material

Download Coleman et al. supplementary material(File)
File 70.4 KB