Next Article in Journal
Improving Bimonthly Landscape Monitoring in Morocco, North Africa, by Integrating Machine Learning with GRASS GIS
Previous Article in Journal
Mapping Spatial Variability of Sugarcane Foliar Nitrogen, Phosphorus, Potassium and Chlorophyll Concentrations Using Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Examining Deep Learning Pixel-Based Classification Algorithms for Mapping Weed Canopy Cover in Wheat Production Using Drone Data

Department of Agricultural Science and Engineering, College of Agriculture, Tennessee State University, Nashville, TN 37209, USA
*
Author to whom correspondence should be addressed.
Submission received: 17 December 2024 / Revised: 7 January 2025 / Accepted: 8 January 2025 / Published: 10 January 2025

Abstract

:
Deep learning models offer valuable insights by leveraging large datasets, enabling precise and strategic decision-making essential for modern agriculture. Despite their potential, limited research has focused on the performance of pixel-based deep learning algorithms for detecting and mapping weed canopy cover. This study aims to evaluate the effectiveness of three neural network architectures—U-Net, DeepLabV3 (DLV3), and pyramid scene parsing network (PSPNet)—in mapping weed canopy cover in winter wheat. Drone data collected at the jointing and booting growth stages of winter wheat were used for the analysis. A supervised deep learning pixel classification methodology was adopted, and the models were tested on broadleaved weed species, winter wheat, and other weed species. The results show that PSPNet outperformed both U-Net and DLV3 in classification performance, with PSPNet achieving the highest overall mapping accuracy of 80%, followed by U-Net at 75% and DLV3 at 56.5%. These findings highlight the potential of pixel-based deep learning algorithms to enhance weed canopy mapping, enabling farmers to make more informed, site-specific weed management decisions, ultimately improving production and promoting sustainable agricultural practices.

1. Introduction

Weed infestation remains a major problem in wheat production [1]. Weeds often grow contemporaneously, through different growth stages in the phenology of winter wheat. The occurrence of certain weed species depends largely on soil properties and environmental conditions, such as climate and management practices [2]. These conditions allow for the recurring appearance of weeds on many winter wheat fields. The concept of manual weed scouting is generally marked by time and labor constraints, little efficacy, close monitoring, and lots of effort from farmers, especially in large agricultural fields [3,4].
Drones have become effective platforms for obtaining field information for both agricultural and non-agricultural activities [5]. It provides flexibility of flight altitudes and high-spatial-resolution imagery (≥1 cm) to support crops and weed mapping [6]. The continuous effort by farmers to improve crop yield and maximize agricultural productivity is greatly enhanced by drone and satellite technologies [7,8]. Drones also offer farmers the possibility of close monitoring and early-stage detection of weeds in agricultural production such as winter wheat [9,10]. The combined use of drone imagery and remote sensing techniques has adequately proven to be reliable in obtaining growth indices across every stage of crop production [11,12].
Deep learning (DL) is a component of machine learning and, to a larger extent, artificial intelligence (AI). Deep learning uses neural networks (NN) to detect image objects or for image classification [13]. It is a data-driven approach that enables more precise and strategic decision-making [14]. Automated weed canopy cover mapping and analysis through deep learning approaches reduce the need for manual labor and traditional surveying methods, both of which are often time-consuming and labor-intensive [15]. Deep learning classification techniques have been commonly used in areas such as computer vision and pattern recognition [16,17]. However, applications in the agricultural sector are still emerging. Efficient processing of both multispectral and hyperspectral remotely sensed data has been reported using deep learning techniques [18,19,20].
For site-specific weed mapping, de Camargo et al. [17] proposed an optimized deep learning (DL) approach that achieved high accuracy (94%) when differentiating crops and weeds using high-resolution images acquired from unmanned aerial vehicles (UAVs). To understand how traditional image classifiers and deep learning classification algorithms perform, Kussul et al. [19] used the random forest classifier and deep learning (DL) models to map land cover and crop types. They found higher overall accuracy rates for the DL models (92.7–94.6%) compared to the random forest classifier (88.7%). Despite their high performance in image segmentation, Tao et al. [13] assert that changes in the depth of input images, the quality of training data, and model overfitting or underfitting may affect the final output and overall accuracy of DL model architectures used for image pixel classification.
While previous studies have employed deep learning techniques for mapping crop and weed types, there remains a gap in research exploring the comparative performance of different pixel-based deep learning algorithms specifically for mapping weed canopy cover in winter wheat production. Most existing studies have focused on broader applications of deep learning in agriculture or have evaluated individual models without delving into their relative strengths and weaknesses in specific contexts, such as weed mapping in high-density crop systems. Recognizing this gap, the aim of this study is to evaluate and compare the performance of U-Net, DeepLabV3, and pyramid scene parsing network (PSPNet) model classifiers in accurately mapping weed canopy cover within a winter wheat field, using drone-acquired imagery. By doing so, this research seeks to build upon the findings of prior studies, providing a more detailed understanding of the potential and limitations of these models in supporting precision and sustainable agriculture.

2. Materials and Methods

2.1. Study Area

The study was conducted on the Tennessee State University’s urban agricultural field, located in Davidson County (Figure 1). The part of the field used for the study lies between latitude 36.176° N and longitude 86.827° W, close to the Cumberland River. Located in the southeastern region of the country, the area experiences cold winters and warm summers. The average annual high temperature of the area is about 70 °F (21.1 °C) and the average annual low temperature is about 49 °F (9.4 °C). The mean annual precipitation is around 47.2 inches. The soil type of the area is predominantly Byler silt loam (ByB), a moderately acidic soil formed from weathered limestone materials [21].

2.2. Methodology

The methodology (Figure 2) used involved the growing of an unknown winter wheat species. At four growing stages of the wheat (tillering, jointing, booting, and maturity), multispectral images of the field were taken using a drone. The captured images were labeled (trained to identify the different features) and used to train deep leaning models for the jointing and booting growth stages. The models were then used to classify the input images to differentiate the major broad-leaved weeds from the winter wheat and other weed species in the wheat plots. An assessment of the model architectures was made to determine their robustness.

2.2.1. Growing of Winter Wheat

An unknown winter wheat variety was planted on 19 October 2022, in 6 m × 6 m plots. The field was burnt with a non-selective herbicide (2 percent Roundup®) to manage the initial weeds. The planting of the winter wheat was conducted 2 weeks after the burning with a no-till planter. At the tillering stage, 45 kg of nitrogen was applied to the wheat plots based on recommendations from soil testing. The matured wheat was cut down with a rotary cutter mower on 10 June 2023.

2.2.2. Drone Data Acquisition

An Inspire-2 drone equipped with an Altum multispectral camera was used to take images of the cultivated field during four growth stages (tillering, jointing, booting, and mature) of winter wheat. The drone was flown at an altitude of 15 m above ground level at a speed of 3 m/s, with an overlap rate of 80–90% and a spectral resolution of 1 cm. The images were captured within the spectral bands of blue (450 to 520 nm), green (520 to 590 nm), red (630 to 690 nm), red-edge (690 to 730 nm), near-infrared (NIR) (770–890 nm), and longwave infrared thermal (LWIR) (10,600 to 11,200 nm). The drone images were taken every 2 seconds. The drone images had a spatial resolution of about 1 cm. The images captured were geotagged, radiometrically corrected (using the camera and sun radiation), and orthomosaiced in Pix4D mapper (version 4.8.0). The mosaicked imagery was then clipped to the wheat plots.

2.2.3. Deep Learning Image (Pixel) Classification

A single wheat plot, representative of both the jointing and booting growth stages of the winter wheat, was used for training and classification with the three deep learning model architectures to evaluate their performance in segmenting and distinguishing broad-leaved weed species from winter wheat and other weeds. Supervised classification was performed on the images using ArcGIS Pro (version 3.3.1). The jointing and booting stages were selected for this evaluation due to the dominance and competitiveness of weeds during those stages. A training dataset was created from the input images, with over 2012 and 2205 polygons digitized for the jointing and booting growth stages, respectively. Digitization was based on the distinct floral colors of the weed species visible in the drone imagery (Figure 3) and field digital images. These training data were then exported as classified tiles for deep learning model training with the U-Net, DeepLabV3, and pyramid scene parsing network (PSPNet) architectures, utilizing ResNet34 as the backbone model. The models were trained with 224 × 224 pixel-sized inputs, for a maximum of 20 epochs to ensure consistency. The dataset was split into 80% for model training and 20% for the validation of the model. The two images were classified four times with each model.

2.2.4. Overview of Deep Learning Model Architectures

U-Net, DeepLabV3, and pyramid scene parsing network (PSPNet) are commonly used deep learning architectures for semantic segmentation tasks due to their proven effectiveness in extracting spatial features and accurately segmenting complex image data [22,23,24]. These models were evaluated for their capacity to accurately segment broad-leaved weeds in a winter wheat field. The evaluation aimed to determine how well each model could identify and separate broad-leaved weeds from the surrounding winter wheat crop, a process which is essential for effective weed management and precision agriculture. Each of these models is further explained below.
The U-Net model classifier architecture works within an encoder–decoder workflow [23,25,26]. The encoder consists of several convolutional layers followed by rectified linear unit (ReLU) activations and max-pooling layers. The decoder reconstructs the segmented image from the encoded feature representations. Furthermore, it uses upsampling layers to increase the spatial dimensions of feature maps, thus refining the segmentation. Each upsampling step is followed by convolutional layers to improve the resolution of the output (Figure 4).
The DeepLabV3 architecture (Figure 5) uses atrous convolution (AC) and atrous spatial pyramid pooling (ASPP) methods in its workflow [28,29,30]. The AC is utilized to increase the receptive field of the convolutional layers without losing spatial resolution, allowing the network to capture more contextual information without downsampling the feature maps. ASPP is designed to gather multi-scale contextual information by applying atrous convolutions with different dilation rates in parallel. This enables the model to recognize objects at multiple scales and improves its ability to segment images accurately.
The pyramid scene parsing network classifier architecture is centered on a pyramid pooling module (PPM) of contextual information at multiple scales of the input dataset [24,32]. The PPM consists of several parallel pooling layers with different grid sizes. Each pooling layer partitions the input feature map into different regions and performs pooling within these regions to generate fixed-size feature maps (Figure 6). The enriched feature representation from the PPM is fed into a series of convolutional layers that produce the segmented image. Each pixel in the output map is assigned a class label, corresponding to the different objects or features within the input image [33].

2.2.5. Assessment of the Models

The classified weed canopy cover maps derived using the algorithms from the three model architectures (U-Net, DLV3, and PSPNet) were validated to assess the accuracy and performance of the deep learning algorithms. The validation and accuracy assessments were carried out using the precision, recall, and overall accuracy criteria. The digital field images, together with 400 (jointing stage) and 500 (booting stage) equalized random points, were used to assess the classification accuracy of the models. The precision was estimated using Equation (1), recall using Equation (2), and overall accuracy using Equation (3) [17].
Precision   ( p ) = T P i T P i + F P i
Recall   ( r ) = T P i T P i + F P i
Overall   Accuracy   ( o a ) = i k = 1 C i N
where:
  • T P i = true positive of class i
  • F P i = false positive of class i
  • o a = overall accuracy
  • C i   = the count of true positives for class i
  • i k C i = the sum of all true positives across all classes
  • N = the total number of instances in the matrix

3. Results

The training loss is the training error of a model as it learns from the training data, while the validation loss demonstrates whether the model underfits or overfits the training dataset. At the jointing stage, the training loss (TL) for the U-Net model started high and decreased steadily as the model learned from the training data, while the validation loss (VL) started low initially and stayed constant as it converged with the TL curve. Similarly, for the DLC3 curve, the training loss started high and decreased steadily with the increase in the number of processed batches. The validation curve started low and stayed constant throughout the processed batches. In contrast, both the training and validation losses in the PSPNet started high and significantly dropped throughout the processed batches (Figure 7).
The pattern of the training and validation losses derived from the U-Net, DLV3, and PSPNet classification models (Figure 8) in the booting stage of winter wheat were like the jointing growth stage loss curves. The training loss in both U-Net and DLV3 started high and sharply decreased with the increase in the number of processed batches. In contrast, the validation loss for both U-Net and DVLV3 started low and steadily plateaued out as the number of processed batches increased. Both training and validation losses in the PSPNet classification algorithms started high and sharply decreased with the increase in the number of processed batches.
Figure 9 compares the performance of U-Net, pyramid scene parsing network (PSPNet), and DeepLabV3 (DLV3) in weed mapping during the jointing growth stage of winter wheat. Both U-Net and PSPNet effectively captured the overall weed distribution, though they tended to overclassify the speedwell species. DeepLabV3 delivered the most refined segmentation, particularly in relation to the differentiation of speedwell and mayweed from winter wheat and other weed species. Generally, U-Net and PSPNet produced more detailed classification maps, while DLV3 provided a balance between clarity and broader weed detection.
Figure 10 illustrates the performance of the three deep learning model classifiers in mapping weeds during the booting growth stage of winter wheat. The U-Net model captured a broad distribution of weeds but showed a notable underclassification, particularly in relation to common vetch and the other weed species. PSPNet produced a more intricate segmentation, though it blended multiple weed species, resulting in a denser yet less distinct classification. In contrast, DLV3 delivered a more distinct segmentation, isolating species like mayweed and hairy buttercup while missing the other weed species. Overall, DLV3 offered the most precise weed identification, while U-Net and PSPNet balanced broader weed detection with varying levels of detail and clarity.
The accuracy assessments in Table 1, Table 2 and Table 3 for the jointing stage of winter wheat highlight the classification performance of the models across four vegetation classes. U-Net demonstrated the highest overall accuracy at 81% (Table 1), with the highest precision for mayweed (98%) and the lowest for speedwell (73%). DLV3 recorded the lowest overall accuracy at 65%, with mayweed achieving the highest precision (71%) and the other species having the lowest (56%). PSPNet, with an overall accuracy of 77%, delivered strong results with a precision of 91% for mayweed and 71% for winter wheat (Table 3), showing a competitive performance across key metrics.
At the booting growth stage, the U-Net model achieved an overall accuracy of 69%, with mayweed having the highest precision (90%) and the smallest precision attained for other weed species at 80% (Table 4). The DLV3 model recorded an overall accuracy of 48%, with notable precision for mayweed (91%) and no precision or recall for common vetch and other species (Table 5). PSPNet outperformed the other classifiers with an overall accuracy of 82% (Table 6), showing strong precision across weeds, particularly for hairy buttercup (88%).

4. Discussion

In this paper, we examined the performance of three deep learning models—U-Net, DLV3, and PSPNet—in mapping weed canopy cover within a winter wheat field using drone-acquired images. The training and validation loss (TL and VL) curves revealed key factors that influenced the performance of the three models. At the jointing stage, U-Net’s rapid decline in TL and early VL stabilization indicated efficient learning with minimal overfitting, contributing to its high mapping accuracy. DLV3 showed closely aligned TL and VL, reflecting a balanced training and effective generalization. PSPNet demonstrated a sharp initial learning and a steady loss reduction, supporting its balanced detection capabilities. At the booting stage, U-Net maintained low and stable TL and VL, ensuring a consistent performance, while DLV3’s minimal TL-VL gap highlighted its reliability. PSPNet’s gradual and aligned loss reductions indicated robust generalization, aiding its ability to handle complex vegetation patterns. These trends underscore the potential of these models for effective weed detection and management based on their specific strengths [34].
At the jointing growth stage, all three models demonstrated unique strengths and limitations in mapping weeds among the winter wheat. Both the U-Net and PSPNet classifiers effectively captured the overall weed layout but struggled with oversegmentation, particularly when identifying speedwell. DeepLabV3 (DLV3), on the other hand, excelled at producing detailed and refined segmentation, accurately differentiating speedwell and mayweed from winter wheat and other weed species. During the booting growth stage, U-Net provided a broad weed coverage but suffered from underclassification, particularly for common vetch and other weeds, missing critical details. PSPNet delivered a more intricate weed map but tended to blend multiple species, resulting in overly dense and less distinct classifications. DLV3, again, demonstrated its strength in precision, isolating mayweed and hairy buttercup effectively, although it failed to detect some other weed species, indicating gaps in comprehensive weed detection. In general, the outputs from DLV3 are more applicable for targeted weed control strategies, while the outputs from both PSPNet and U-Net, with their broader weed mapping capabilities, are more suitable for individual weed identification.
The varying classification performance of the three models can be attributed to their distinct architectural designs and to how they process spatial features and contextual information. U-Net’s symmetric encoder–decoder architecture is highly effective at capturing fine-grained details and spatial relationships [35], enabling it to distinguish individual weed species such as speedwell and hairy buttercup with notable accuracy. However, the model’s reverse learning process may have contributed to the missed classifications observed during the booting stage [36]. PSPNet, with its pyramid pooling module, effectively captures multi-scale contextual information [37], enabling a balanced performance across feature segmentation and detection [38]. This ability to detect all weed species at both growth stages, though with noticeable blending, can be linked to its emphasis on contextual understanding [39]. In contrast, DeepLabV3 (DLV3) utilizes atrous convolution layers for dense spatial sampling, which, while enhancing broader feature detection, struggles with capturing finer details [40]. This architectural choice likely contributed to its reduced accuracy in identifying smaller or less distinct weeds, leading to occasional misclassifications. The quality of the input image, as well as that of the training dataset, could have also contributed to the variations in the model’s performance.
When evaluating robustness, U-Net achieved an overall accuracy of 75%, indicating a strong ability to generalize weed mapping. This generalization makes it suitable for large-scale applications but less effective in scenarios where high precision is critical [41,42]. PSPNet, with the highest overall accuracy of 80%, demonstrated superior resilience by balancing clarity and weed identification, making it well-suited for mapping complex vegetation patterns [43]. In contrast, DeepLabV3 (DLV3) achieved an overall accuracy of 56.5%, showing a vulnerability to underclassification which limits its effectiveness in highly variable environments [44]. Despite its lower overall accuracy, DLV3 excelled at precision, making it particularly effective for tasks requiring detailed, species-specific weed mapping. PSPNet’s detailed outputs provide an advantage for mapping diverse weed environments but may struggle with species blending [45], which can reduce its robustness in certain applications. U-Net, while reliable for broader weed coverage, may face challenges in detecting less dominant species, affecting its utility in scenarios requiring detailed mapping. Overall, these results suggest that U-Net and PSPNet are best suited for general weed mapping with high accuracy, while DLV3’s precision makes it ideal for focused weed management and species-specific strategies.
The study demonstrates that using high-resolution drone imagery combined with effective deep learning models can significantly enhance weed management practices, promoting sustainability by reducing blanket herbicide applications. Understanding the relative abundance and distribution of weeds necessitates site-specific weed management practices. Precision agriculture techniques, such as variable rate herbicide application, could then be employed to target weed hotspots more effectively. Overall, this study provides valuable insights into the effectiveness of different deep learning models in the context of precision agriculture, with PSPNet emerging as a particularly strong candidate for accurate weed mapping. Future research should explore the impact of different soil tillage systems on weed cover dynamics during winter wheat production.

5. Conclusions

This study sought to evaluate and compare the performance of U-Net, DeepLabV3 (DLV3), and pyramid scene parsing network (PSPNet) classifiers in accurately mapping weed canopy cover within a winter wheat field using drone imagery. PSPNet emerged as the most accurate model with an overall accuracy of 80%, excelling at general weed mapping across complex vegetation patterns. U-Net achieved a 75% accuracy, demonstrating strong generalization suitable for large-scale weed mapping, while DLV3, with an accuracy of 56.5%, provided high precision for species-specific weed identification. The performance of these models reflects their architectural strengths, with PSPNet balancing clarity and detection, U-Net capturing fine details but struggling with underclassification, and DLV3 excelling at detailed segmentation but missing finer distinctions. These findings underscore the potential of deep learning models to enhance site-specific weed management practices, promoting sustainability in agriculture. Future research will focus on the development of an enhanced DLV3 model tailored for distinguishing different weed types, with the aim of generating precise prescription maps for targeted weed control.

Author Contributions

J.N.O.: conceptualization, methodology, data curation, formal analysis, visualization, writing—original draft, validation; C.E.A.: writing—review and editing, supervision, project administration; S.D.: writing—review and editing, supervision; S.A.: data acquisition and curation. All authors have read and agreed to the published version of the manuscript.

Funding

Funded by the United States Department of Agriculture (USDA)-National Institute of Food and Agriculture (NIFA) through the Agriculture and Food Research Initiative (AFRI) Small and Medium-Sized Farms program, grant number 2021-69006-33875. Project director: Akumu Clement E.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available upon request.

Conflicts of Interest

There are no conflicts of interest.

References

  1. Flessner, M.L.; Burke, I.C.; Dille, J.A.; Everman, W.J.; VanGessel, M.J.; Tidemann, B.; Manuchehri, M.R.; Soltani, N.; Sikkema, P.H. Potential wheat yield loss due to weeds in the United States and Canada. Weed Technol. 2021, 35, 916–923. [Google Scholar] [CrossRef]
  2. Santín-Montanyá, M.I.; Martín-Lammerding, D.; Walter, I.; Zambrana, E.; Tenorio, J.L. Effects of tillage, crop systems and fertilization on weed abundance and diversity in 4-year dry land winter wheat. Eur. J. Agron. 2013, 48, 43–49. [Google Scholar] [CrossRef]
  3. Hall, D.; Dayoub, F.; Kulk, J.; McCool, C. Towards unsupervised weed scouting for agricultural robotics. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5223–5230. [Google Scholar]
  4. Kalischuk, M.; Paret, M.L.; Freeman, J.H.; Raj, D.; Da Silva, S.; Eubanks, S.; Wiggins, D.; Lollar, M.; Marois, J.J.; Mellinger, H.C. An improved crop scouting technique incorporating unmanned aerial vehicle–assisted multispectral crop imaging into conventional scouting practice for gummy stem blight in watermelon. Plant Dis. 2019, 103, 1642–1650. [Google Scholar] [CrossRef] [PubMed]
  5. Mateen, A.; Zhu, Q. Weed Detection in Wheat Crop Using UAV for Precision Agriculture. Pak. J. Agric. Sci. 2017, 56, 775–784. [Google Scholar] [CrossRef]
  6. Mohd Noor, N.; Abdullah, A.; Hashim, M. Remote sensing UAV/drones and its applications for urban areas: A review. IOP Conf. Ser. Earth Environ. Sci. 2018, 169, 012003. [Google Scholar] [CrossRef]
  7. Pflanz, M.; Nordmeyer, H.; Schirrmann, M. Weed Mapping with UAS Imagery and a Bag of Visual Words Based Image Classifier. Remote Sens. 2018, 10, 1530. [Google Scholar] [CrossRef]
  8. Abiri, R.; Rizan, N.; Balasundram, S.K.; Shahbazi, A.B.; Abdul-Hamid, H. Application of digital technologies for ensuring agricultural productivity. Heliyon 2023, 9, e22601. [Google Scholar] [CrossRef]
  9. Gómez-Candón, D.; De Castro, A.I.; López-Granados, F. Assessing the accuracy of mosaics from unmanned aerial vehicle (UAV) imagery for precision agriculture purposes in wheat. Precis. Agric. 2013, 15, 44–56. [Google Scholar] [CrossRef]
  10. Kenawy, E.-S.M.; Khodadadi, N.; Mirjalili, S.; Makarovskikh, T.; Abotaleb, M.; Karim, F.K.; Alkahtani, H.K.; Abdelhamid, A.A.; Eid, M.M.; Horiuchi, T. Metaheuristic optimization for improving weed detection in wheat images captured by drones. Mathematics 2022, 10, 4421. [Google Scholar] [CrossRef]
  11. Yang, C. Remote Sensing and Precision Agriculture Technologies for Crop Disease Detection and Management with a Practical Application Example. Engineering 2020, 6, 528–532. [Google Scholar] [CrossRef]
  12. Fu, Z.; Jiang, J.; Gao, Y.; Krienke, B.; Wang, M.; Zhong, K.; Cao, Q.; Tian, Y.; Zhu, Y.; Cao, W. Wheat growth monitoring and yield estimation based on multi-rotor unmanned aerial vehicle. Remote Sens. 2020, 12, 508. [Google Scholar] [CrossRef]
  13. Tao, Y.; Xu, M.; Lu, Z.; Zhong, Y. DenseNet-Based Depth-Width Double Reinforced Deep Learning Neural Network for High-Resolution Remote Sensing Image Per-Pixel Classification. Remote Sens. 2018, 10, 779. [Google Scholar] [CrossRef]
  14. Zheng, Y.-Y.; Kong, J.-L.; Jin, X.-B.; Wang, X.-Y.; Su, T.-L.; Zuo, M. CropDeep: The Crop Vision Dataset for Deep-Learning-Based Classification and Detection in Precision Agriculture. Sensors 2019, 19, 1058. [Google Scholar] [CrossRef]
  15. Nguyen, T.T.; Hoang, T.D.; Pham, M.T.; Vu, T.T.; Nguyen, T.H.; Huynh, Q.-T.; Jo, J. Monitoring agriculture areas with satellite images and deep learning. Appl. Soft Comput. 2020, 95, 106565. [Google Scholar] [CrossRef]
  16. Busia, A.; Dahl, G.E.; Fannjiang, C.; Alexander, D.H.; Dorfman, E.; Poplin, R.; McLean, C.Y.; Chang, P.-C.; DePristo, M. A deep learning approach to pattern recognition for short DNA sequences. bioRxiv 2019. [Google Scholar] [CrossRef]
  17. de Camargo, T.; Schirrmann, M.; Landwehr, N.; Dammer, K.-H.; Pflanz, M. Optimized Deep Learning Model as a Basis for Fast UAV Mapping of Weed Species in Winter Wheat Crops. Remote Sens. 2021, 13, 1704. [Google Scholar] [CrossRef]
  18. Yang, X.; Ye, Y.; Li, X.; Lau, R.Y.K.; Zhang, X.; Huang, X. Hyperspectral Image Classification With Deep Learning Models. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5408–5423. [Google Scholar] [CrossRef]
  19. Kussul, N.; Lavreniuk, M.; Skakun, S.; Shelestov, A. Deep Learning Classification of Land Cover and Crop Types Using Remote Sensing Data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 778–782. [Google Scholar] [CrossRef]
  20. Singh, M.; Tyagi, K.D. Pixel based classification for Landsat 8 OLI multispectral satellite images using deep learning neural network. Remote Sens. Appl. Soc. Environ. 2021, 24, 100645. [Google Scholar] [CrossRef]
  21. Akumu, C.E.; Oppong, J.N.; Dennis, S. Examining the Percent Canopy Cover and Health of Winter Wheat in No-Till and Conventional Tillage Plots Using a Drone. Agriculture 2024, 14, 760. [Google Scholar] [CrossRef]
  22. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  23. Diao, Z.; Guo, P.; Zhang, B.; Zhang, D.; Yan, J.; He, Z.; Zhao, S.; Zhao, C. Maize crop row recognition algorithm based on improved UNet network. Comput. Electron. Agric. 2023, 210, 107940. [Google Scholar] [CrossRef]
  24. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  25. Taparia, A. U-Net Architecture Explained. Available online: https://www.geeksforgeeks.org/u-net-architecture-explained/# (accessed on 11 June 2024).
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  27. ArcGIS Developers. How U-Net Works? For ESRI. Available online: https://developers.arcgis.com/python/guide/how-unet-works/ (accessed on 11 June 2024).
  28. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  29. Liu, M.; Fu, B.; Xie, S.; He, H.; Lan, F.; Li, Y.; Lou, P.; Fan, D. Comparison of multi-source satellite images for classifying marsh vegetation using DeepLabV3 Plus deep learning algorithm. Ecol. Indic. 2021, 125, 107562. [Google Scholar] [CrossRef]
  30. Quan, B.; Liu, B.; Fu, D.; Chen, H.; Liu, X. Improved Deeplabv3 For Better Road Segmentation In Remote Sensing Images. In Proceedings of the 2021 International Conference on Computer Engineering and Artificial Intelligence (ICCEAI), Shanghai, China, 27–29 August 2021; pp. 331–334. [Google Scholar]
  31. ArcGIS Developers. How DeepLabV3 Works? For ESRI. Available online: https://developers.arcgis.com/python/guide/how-deeplabv3-works/ (accessed on 11 June 2024).
  32. Long, X.; Zhang, W.; Zhao, B. PSPNet-SLAM: A Semantic SLAM Detect Dynamic Object by Pyramid Scene Parsing Network. IEEE Access 2020, 8, 214685–214695. [Google Scholar] [CrossRef]
  33. ArcGIS Developers. How PSPNet Works? For ESRI. Available online: https://developers.arcgis.com/python/guide/how-pspnet-works/ (accessed on 11 June 2024).
  34. Li, Y.; Zhang, H.; Xue, X.; Jiang, Y.; Shen, Q. Deep learning for remote sensing image classification: A survey. WIREs Data Min. Knowl. Discov. 2018, 8, e1264. [Google Scholar] [CrossRef]
  35. Zhao, X.; Yuan, Y.; Song, M.; Ding, Y.; Lin, F.; Liang, D.; Zhang, D. Use of unmanned aerial vehicle imagery and deep learning unet to extract rice lodging. Sensors 2019, 19, 3859. [Google Scholar] [CrossRef] [PubMed]
  36. Kim, J.; Song, Y.; Lee, W.-K. Accuracy analysis of multi-series phenological landcover classification using U-Net-based deep learning model-Focusing on the Seoul, Republic of Korea. Korean J. Remote Sens. 2021, 37, 409–418. [Google Scholar] [CrossRef]
  37. Zhao, Z.; Liu, X.; Li, M.; Liu, J.; Wang, Z. Oral Microbe Community and Pyramid Scene Parsing Network-based Periodontitis Risk Prediction. Int. Dent. J. 2024. online ahead of print. [Google Scholar] [CrossRef] [PubMed]
  38. Chen, S.; Song, Y.; Su, J.; Fang, Y.; Shen, L.; Mi, Z.; Su, B. Segmentation of field grape bunches via an improved pyramid scene parsing network. Int. J. Agric. Biol. Eng. 2021, 14, 185–194. [Google Scholar] [CrossRef]
  39. Zhang, R.; Chen, J.; Feng, L.; Li, S.; Yang, W.; Guo, D. A Refined Pyramid Scene Parsing Network for Polarimetric SAR Image Semantic Segmentation in Agricultural Areas. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  40. Fu, Y.; Fan, J.; Xing, S.; Wang, Z.; Jing, F.; Tan, M. Image segmentation of cabin assembly scene based on improved RGB-D mask R-CNN. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  41. Naik, K.J. Deep Learning Based Segmentation of Weed Images in crop fields using U-Net. In Proceedings of the 2024 3rd International Conference for Advancement in Technology (ICONAT), Goa, India, 6–8 September 2024; pp. 1–6. [Google Scholar]
  42. Zou, K.; Chen, X.; Zhang, F.; Zhou, H.; Zhang, C. A field weed density evaluation method based on uav imaging and modified u-net. Remote Sens. 2021, 13, 310. [Google Scholar] [CrossRef]
  43. Shen, Y.; Sun, X.; Cui, J.; Lu, Y. Application of Pyramid Scene Parsing Network in leaf segmentation for Wheat Stripe Rust. In Proceedings of the 2024 5th International Conference on Computer Vision, Image and Deep Learning (CVIDL), Zhuhai, China, 19–21 April 2024; pp. 926–930. [Google Scholar]
  44. Zhang, K.; Li, L.; Liu, H.; Yuan, J.; Tai, X.-C. Deep Convolutional Neural Networks Meet Variational Shape Compactness Priors for Image Segmentation. arXiv 2024, arXiv:2406.19400. [Google Scholar] [CrossRef]
  45. Veeragandham, S.; Santhi, H. Optimization enabled Deep Quantum Neural Network for weed classification and density estimation. Expert Syst. Appl. 2024, 243, 122679. [Google Scholar] [CrossRef]
Figure 1. Location of the study area with an insert of the study field.
Figure 1. Location of the study area with an insert of the study field.
Geomatics 05 00004 g001
Figure 2. Schematic representation of the methodology used for mapping weed canopy cover.
Figure 2. Schematic representation of the methodology used for mapping weed canopy cover.
Geomatics 05 00004 g002
Figure 3. Training on multispectral image for DL models.
Figure 3. Training on multispectral image for DL models.
Geomatics 05 00004 g003
Figure 4. U-Net architecture for image segmentation, adapted from [27] (p. 2).
Figure 4. U-Net architecture for image segmentation, adapted from [27] (p. 2).
Geomatics 05 00004 g004
Figure 5. Model architecture for DeepLabV3, adapted from [31] (p. 1).
Figure 5. Model architecture for DeepLabV3, adapted from [31] (p. 1).
Geomatics 05 00004 g005
Figure 6. The PSPNet model architecture, adopted from [25] (p. 2884).
Figure 6. The PSPNet model architecture, adopted from [25] (p. 2884).
Geomatics 05 00004 g006
Figure 7. Jointing stage training and validation loss graphs.
Figure 7. Jointing stage training and validation loss graphs.
Geomatics 05 00004 g007
Figure 8. Booting growth stage training and validation loss graphs.
Figure 8. Booting growth stage training and validation loss graphs.
Geomatics 05 00004 g008
Figure 9. Classified weed canopy cover map derived from the three model classifiers during the jointing growth stage.
Figure 9. Classified weed canopy cover map derived from the three model classifiers during the jointing growth stage.
Geomatics 05 00004 g009
Figure 10. Classified weed canopy cover map derived from the three model classifiers during the booting growth stage.
Figure 10. Classified weed canopy cover map derived from the three model classifiers during the booting growth stage.
Geomatics 05 00004 g010
Table 1. Accuracy assessment metrics for weed canopy cover derived using the U-Net classifier at the jointing stage of winter wheat.
Table 1. Accuracy assessment metrics for weed canopy cover derived using the U-Net classifier at the jointing stage of winter wheat.
U-Net—Jointing
Reference/classMayweedSpeedwellOthersWheat
Mayweed6524101
Speedwell19251
Others087616
Wheat02591
p (%)98737983
r (%)65927691
OA (%)81
Table 2. Accuracy assessment metrics for weed canopy cover derived using the DLV3 classifier at the jointing stage of winter wheat.
Table 2. Accuracy assessment metrics for weed canopy cover derived using the DLV3 classifier at the jointing stage of winter wheat.
DVL3—Jointing
Reference/classMayweedSpeedwellOthersWheat
Mayweed7012153
Speedwell2255158
Others5135527
Wheat041381
p (%)71655668
r (%)70555581
OA (%)65
Table 3. Accuracy assessment metrics for weed canopy cover derived using the PSPNet classifier at the jointing stage of winter wheat.
Table 3. Accuracy assessment metrics for weed canopy cover derived using the PSPNet classifier at the jointing stage of winter wheat.
PSPNet—Jointing
Reference/classMayweedSpeedwellOthersWheat
Mayweed632872
Speedwell18807
Others127126
Wheat43687
p (%)0.910.730.810.71
r (%)0.630.880.710.87
OA (%)77
Table 4. Accuracy assessment metrics for weed canopy cover derived using the U-Net classifier at the booting stage of winter wheat.
Table 4. Accuracy assessment metrics for weed canopy cover derived using the U-Net classifier at the booting stage of winter wheat.
U-Net—Booting
Reference/classMayweedHairy buttercupCommon vetchOthersWheat
Mayweed866026
Hairy buttercup686053
Common vetch020296
Others3008314
Wheat120592
p (%)909008644
r (%)868608392
OA (%)69
Table 5. Accuracy assessment metrics for weed canopy cover derived using the DLV3 classifier at the booting stage of winter wheat.
Table 5. Accuracy assessment metrics for weed canopy cover derived using the DLV3 classifier at the booting stage of winter wheat.
DVL3—Booting
Reference/classMayweedHairy buttercupCommon vetchOthersWheat
Mayweed62170021
Hairy buttercup2790019
Common vetch640096
Others130096
Wheat120097
p (%)91770029
r (%)62790097
OA (%)48
Table 6. Accuracy assessment metrics for weed canopy cover derived using the PSPNet classifier at the booting stage of winter wheat.
Table 6. Accuracy assessment metrics for weed canopy cover derived using the PSPNet classifier at the booting stage of winter wheat.
PSPNet—Booting
Reference/classMayweedHairy buttercupCommon vetchOthersWheat
Mayweed7636510
Hairy buttercup384355
Common vetch238345
Others3428110
Wheat634780
p (%)8488867875
r (%)7684838080
OA (%)82
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Oppong, J.N.; Akumu, C.E.; Dennis, S.; Anyanwu, S. Examining Deep Learning Pixel-Based Classification Algorithms for Mapping Weed Canopy Cover in Wheat Production Using Drone Data. Geomatics 2025, 5, 4. https://doi.org/10.3390/geomatics5010004

AMA Style

Oppong JN, Akumu CE, Dennis S, Anyanwu S. Examining Deep Learning Pixel-Based Classification Algorithms for Mapping Weed Canopy Cover in Wheat Production Using Drone Data. Geomatics. 2025; 5(1):4. https://doi.org/10.3390/geomatics5010004

Chicago/Turabian Style

Oppong, Judith N., Clement E. Akumu, Samuel Dennis, and Stephanie Anyanwu. 2025. "Examining Deep Learning Pixel-Based Classification Algorithms for Mapping Weed Canopy Cover in Wheat Production Using Drone Data" Geomatics 5, no. 1: 4. https://doi.org/10.3390/geomatics5010004

APA Style

Oppong, J. N., Akumu, C. E., Dennis, S., & Anyanwu, S. (2025). Examining Deep Learning Pixel-Based Classification Algorithms for Mapping Weed Canopy Cover in Wheat Production Using Drone Data. Geomatics, 5(1), 4. https://doi.org/10.3390/geomatics5010004

Article Metrics

Back to TopTop