Open Access
31 May 2023 Assessment of deep learning methods for classification of cereal crop growth stage pre and post canopy closure
Sanaz Rasti, Chris J. Bleakley, Guénolé C. M. Silvestre, Gregory M. P. O’Hare, David Langton
Author Affiliations +
Abstract

Growth stage (GS) is an important crop growth metric commonly used in commercial farms. We focus on wheat and barley GS classification based on in-field proximal images using convolutional neural networks (ConvNets). For comparison purposes, use of a conventional machine learning algorithm was also investigated. The research includes extensive data collection of images of wheat and barley crops over a 3-year period. During data collection, videos were recorded during field walks at two camera views: downward looking and 45 deg angled. The resulting dataset contains 110,000 images of wheat and 106,000 of barley taken over 34 and 33 GS classes, respectively. Three methods were investigated as candidate technologies for the problem of GS classification. These methods were: (I) feature extraction and support vector machine, (II) ConvNet with learning from scratch, and (III) ConvNet with transfer learning. The methods were assessed for classification accuracy using test images taken (a) in fields on days imagined in the training data (i.e., seen field-days GS classification) and (b) in fields on days not imagined in the training data (i.e., unseen field-days principal GS classification). Of the three methods investigated, method III achieved the best accuracy for both classification tasks. The model achieved 97.3% and 97.5% GS classification accuracy for seen field-day test data for wheat and barley, respectively. The model also achieved accuracies of 93.5% and 92.2% for the principal GS classification task for wheat and barley, respectively. We provide a number of key research contributions: the collection curation and exposure of a unique GS labeled proximal image dataset of wheat and barley crops, GS classification, and principal GS classification of cereal crops using three different machine learning methods as well as a comprehensive evaluation and comparison of the obtained results.

1.

Introduction

It is projected that in the period 2005 to 2050, food production must increase by 100% to 110% to meet rising demand due to population growth.1 Moreover, there is increasing pressure on producers to reduce the area of land cleared and the cost of food production. As a result, there is a need for improved crop production management and more efficient utilization of resources.

Enhanced food production requires better decision making for crop husbandry and automated crop growth monitoring. Remote sensing can provide useful data on crop growth at the sub-field level. However, remote sensing is currently limited in terms of spatial and temporal accuracy, particularly in regions that are often cloudy.2 Recently, the use of in-field proximal images coupled with computer vision3 techniques has shown promise for automatic crop growth monitoring.4

Growth stage (GS) is a key metric for quantifying cereal crop growth in production fields.5 GS indicates the development stage of the crop by means of a predefined numeric scale, such as Agriculture and Horticulture Development Board (AHDB),6 Zadoks,7 or Biologische Bundesanstalt, Bundessortenamt and CHemical (BBCH).8 The ability to routinely estimate the GS provides crucial input into crop growth models and help inform novel crop husbandry practices. Typically, GS is determined in the field by means of visual inspection by an agricultural scientist (agronomist), or operator, who has sufficient knowledge of GS metrics.

Cereal crop GS estimation can benefit from the application of image processing techniques in a number of ways. First, image data could be recorded at low cost without damaging the crops. Field GS surveys could be collected by cameras affixed to vehicles traversing the field for the purposes of input application,9 by low flying drones,10 or by ground-based robots.11 A point GS estimate could be obtained from a smartphone. This GS information can then be utilized by the farmer for decision making in regard to field inputs.

The research reported herein addresses the problem of estimation of cereal crop GS based on in-field proximal images. The study investigates the use of machine learning algorithms for GS classification of wheat and barley from images. The work focuses on images that are collected from wheat and barley crops in downward and 45-deg-angled looking mode at a height of around 2m above the ground. Ground truth data are collected at field level and labeled using the Zadoks GS scale metric.7 Due to the visual complexity of crop images and growth development stages in cereals, GS estimation by means of image processing is a challenging research problem.12 Moreover, variations in seed rate, crop variety, soil density and dynamic weather conditions such as wind or changes in natural lighting add to the difficulty of GS estimation from images.

For this study, data were collected from fields in Ireland. Wheat image data were collected for two cultivars of Costello and JB Diego winter wheat during their growing season from early October to mid-August. Barley data were collected from two cultivars of Cassia and Infinity winter barley during their growing season from early November to end of July. Image data with frost or unwanted objects/particles that visually occluded the crops were manually removed from the dataset.

To the best of the authors’ knowledge, this is the first paper to investigate GS classification of cereal crops for a wide range of GSs, including images pre and post canopy closure. In addition, the paper investigates classification accuracy of excluding test images taken on the same day and same field as training images. Herein, we refer to this as testing of unseen field-day data. Due to limitations in the number of GSs in the dataset, the GSs are classified into principal GS, rather than individual GS. In other words, the GSs are grouped into principal GS as classes. The impact of employing principal GSs and lack of images with matching GSs in training and test sets on classification accuracy is studied. Various experiments were carried out and the outcome of best performing algorithm is investigated for GS classification and principal GS classification of downward and 45-deg-angled looking images.

The remainder of this paper is structured as follows. Section 2 presents the background and existing approaches to the problem. Section 3 presents details of the collected image dataset for wheat and barley. The experimental methods and results are presented in Sec. 4. A comprehensive discussion and the conclusions of the work are presented in Secs. 5 and 6, respectively.

2.

Background and Existing Work

A comprehensive survey in Ref. 4 presents image processing techniques reported in the literature for extracting key cereal crop growth metrics from proximal images. One of the dominant crop growth metrics is cereal GS. To date, little research has been done on automated estimation of GS.

An automated image-based scheme was proposed by12 to detect two principal GSs of corn: emergence and three-leaf stage. The study involved a small number of training samples and employed an image segmentation method combined with affinity propagation clustering for classification. The work achieved a classification accuracy of 96.68% for classifying two GSs.

A study reported in Ref. 13 investigated estimation of two distinct GSs of six wheat cultivars. The authors employed scale invariant, low-level feature extraction, mid-level representation (bag-of-visual-words), and a support vector machine (SVM) for classifying two GSs of wheat. Their algorithm achieved on average 91% accuracy.

In a study described in Ref. 14, rice panicles were modeled from 2D rice images. The study targeted mainly one stage of growth when the panicle attributes were developed. Using a morphological operation, the grain area of rice panicles were extracted. The grain weight and the correlation between the grain area and weight parameters were determined. Their algorithm achieved 90% accuracy.

A drone-based approach was proposed in Ref. 15 for classifying four different GSs of rice. The targeted stages were the early phase of rice growth, the vegetative growth phase, the generative growth phase, and the harvest phase. The authors employed a color histogram (leaf color chart feature) and SVM for classifying GS. They achieved 93% accuracy for classifying four different GSs.

Corn sprout GS estimation was investigated in Ref. 16 using red, green, and blue images, over a diminutive period of 6 days growth. The algorithm consisted of cropping the plant region and using a region growing approach as a function of length and time. Moreover, the plant length was measured continuously in real time as ground truth. The authors reported measurement accuracy by comparing the result of image processing to manual measurement counterparts in centimetres. They achieved d=0.2  cm accuracy on average.

A recently published study by the authors of this paper17 presented GS estimation of wheat and barley crops for prior canopy closure stages. The study used 138,000 images from 12 GSs of wheat and 11 GSs of barley in the dataset. The GS classification task was carried out employing three different machine learning methods: (a) a convolutional neural network (ConvNets) model with learning from scratch, (b) a ConvNets model with transfer learning, and (c) conventional SVM classifier. The authors reported classification accuracy of 99.8% on average while using ConvNet with transfer learning method. Although this work was promising, it was limited to GSs of prior canopy closure. Moreover, the classification results achieved for this research were based only on seen field-day data, i.e., the test images were taken on the same days and in the same fields as the training images.

The research reported herein addresses the problem of GS classification for a wide range of GSs. The study reports the result of principal GS classification of unseen field-day test data for both wheat and barley crops; the unseen field-day data are considered as an unbiased test set for the classifier. The highlights, which are achieved for the principal GS classification through an extensive series of experiments, add a remarkable value to the existing literature on automating crop GS classification.

3.

Dataset

The aim of this study is to classify wheat and barley GS using images of crops and state-of-the-art deep neural network models.18 It has been shown that deep neural network models require very large image datasets for training to achieve high accuracy.19 As part of this research, extensive data collection was undertaken for wheat and barley crops. Overall, there are 216,000 images in the dataset from 15 different fields within Ireland. The data collection protocol is presented in Sec. 3.1 and details of the wheat and barley dataset are provided in Secs. 3.2 and 3.3, respectively.

3.1.

Data Collection Protocol

Cereal crop GS is categorized by means of pre-defined scales. Each scale assigns a value to a recognizable crop stage. The most frequently adopted scale is Zadoks.7 The principal GSs of the Zadoks scale are listed in Table 1.

Table 1

Zadoks GS scale metric, including principal and minor growth stages.7

Principal GSMinor GS
00 to 09 Germination00 - Dry seed
01 - Start of water absorption
03 - Seed fully swollen
05 - First root emerged from seed
07 - Coleoptile emerged from seed
09 - First green leaf just at tip of coleoptile
10 to 19 Seedling growth10 - First leaf through coleoptile
11 - First leaf emerged
12 - Two leaves emerged
13 - Three leaves emerged
14 - Four leaves emerged
15 - Five leaves emerged
16 - Six leaves emerged
17 - Seven leaves emerged
18 - Eight leaves emerged
19 - Nine or more leaves emerged
20 to 29 Tillering20 - Main stem only
21 - Main stem and one tiller
22 - Main stem and two tillers
23 - Main stem and three tillers
24 - Main stem and four tillers
25 - Main stem and five tillers
26 - Main stem and six tillers
27 - Main stem and seven tillers
28 - Main stem and eight tillers
29 - Main stem and nine or more tillers
30 to 39 Stem elongation30 - Pseudostem
31 - First node detectable
32 - Second node detectable
33 - Third node detectable
34 - Fourth node detectable
35 - Fifth node detectable
36 - Sixth node detectable
37 - Flag leaf just visible
39 - Flag leaf ligule just visible
40 to 49 Booting41 - Flag leaf sheath extending
43 - Boots just visible swollen
45 - Boots swollen
47 - Flag leaf sheath opening
49 - First awns visible
50 to 59 Ear emergence51 - Tip of ear just visible
53 - Ear quarter emerged
55 - Ear half emerged
57 - Ear three quarters emerged
59 - Ear emergence complete
60 to 69 Anthesis61 - Beginning of anthesis
65 - Anthesis half-way
69 - Anthesis complete
70 to 79 Milk development71 - Kernel water ripe
73 - Early milk
75 - Medium milk
77 - Late milk
80 to 89 Dough development83 - Early dough
85 - Soft dough
87 - Hard dough
90 to 99 Ripening91- Grain hard, difficult to divide
92 - Grain hard, not dented by thumbnail
93 - Grain loosing in daytime
94 - Over-ripe straw dead and collapsing
95 - Seed dormat
96 - Viable seed giving 50% germination
97 - Secondary dormancy induced
99 - Secondary dormancy lost

Ground truth was determined in the field by an agricultural scientist, or operator, who had sufficient knowledge of cereal GS metrics. GS was determined manually by comparing the plants to the objective visual features defined in the scale.

Images were recorded with a DJI Osmo+ camera.20 The DJI Osmo+ includes a camera, gimbal, and a supporting mobile device handle. The recording was captured with 1080 pixel quality and at 30  frames/sec. At each visit, the operator walked the field, along the tramlines for 3 to 6 min recording a video file of crops. Two camera poses were used: vertically downward looking at the field and at a 45-deg declination from the horizon. The camera was held parallel to the sowing rows of the field at a height of 2 m above the ground. In the post-processing stage, the video frames were extracted as image files for training and testing the network. A series of images were extracted and indexed sequentially. To ensure that no two images were the same, frames were extracted with a minimum of 120 ms between each.

The data collected included the ground truth GS, crop cultivar, seed rate, sowing date, date of capture, field global positioning system (GPS), brightness level, and wind speed. The data were captured over 3 years of growing seasons from 2017 to 2019.

3.2.

Wheat Dataset

The seen field-day wheat training dataset consists of 21 GS classes where each class includes 2000 images for training, 600 images for validation, and 1400 images for test purposes. These 21 classes include four classes in the seedling stage, four classes in the tillering stage, two classes in stem elongation, two classes in ear emergence, one in anthesis, three in milk development, two in dough development, and three in ripening. There are overall 84,000 wheat images in this dataset. The wheat training images are from five distinct fields in Ireland and include two different cultivars in Costello and JB Diego. The brightness range in the wheat training dataset varies between 73.0 and 156.2 (AV). There are five different seed rates in the wheat training data. The wind speed at wheat data capture time varied between 6 and 27  km/h. Figure 1 shows sample of wheat images from these two cultivars and various GSs.

Fig. 1

Samples of wheat image from two cultivars of Costello and JB Diego. Image samples of GS14, GS22, GS76, GS23, and GS69 are from downward set of data. Image samples of GS30, GS55, and GS88 are from 45-deg-angled set of data.

JEI_32_3_033014_f001.png

The unseen field-day dataset, which is separate from the training data, consists of 13 GS classes with 2000 images per class. These classes include two from the seedling stage, three from tillering, one from stem elongation, two from ear emergence, two from milk development, one from dough development, and two from ripening. Each class of test data includes 1000 downward and 1000 images of 45-deg-angled looking. Overall, 26,000 images of unseen field-day wheat data are in the test dataset. The data include brightness variation from 75.9 to 168.2 (AV) and three different seed rates. The wind speed at the time of capture wheat test data varied between 16 and 28  km/h. The unseen field-day data are only used for testing, not training.

Details of the wheat dataset are provided in Table 2 and the seen and unseen field-day split is listed in Table 4(a). Information the about fields and their GPS can be found in Table 5.

Table 2

Wheat training and test dataset.

Wheat
IndexGrowth stageCultivarDate capturedFieldDate sowed onSeed-rate (kg/ha)Min/max brightness (AV)Wind (km/h)
111JB DiegoNovember 12, 20191October 9, 2019148.273.0/120.927
212CostelloNovember 1, 20186October 10, 2018156.975.9/134.320
314CostelloJanuary 19, 20182October 19, 2017145.795.2/145.631
415JB DiegoNovember 30, 20187October 14, 2018150.085.6/129.920
516CostelloDecember 10, 20183October 7, 2018156.9104.6/134.09
617CostelloFebruary 21, 20182October 19, 2017145.798.2/144.817
722CostelloFebruary 21, 20193October 7, 2018156.999.2/134.626
823JB DiegoFebruary 22, 20197October 14, 2018150.0103.6/168.217
923JB DiegoFebruary 17, 20201October 9, 2019148.286.3/130.726
1026CostelloApril 20, 20182October 19, 2017145.797.5/143.319
1126JB DiegoApril 12, 20198October 12, 2018164.7101.12/146.516
1228JB DiegoApril 6, 20197October 14, 2018150.0111.2/165.921
1329CostelloMarch 27, 20193October 7, 2018156.995.1/139.56
1430CostelloMay 4, 20182October 19, 2017145.7102.4/145.617
1532JB DiegoMay 10, 20197October 14, 2018150.093.8/147.028
1634CostelloMay 17, 20182October 19, 2017145.798.6/148.318
1752JB DiegoJune 23, 20198October 12, 2018164.789.6/136.025
1855JB DiegoJune 10, 20204October 14, 2019167.395.3/145.621
1957CostelloJune 10, 20196October 10, 2018156.9102.1/149.723
2059JB DiegoMay 14, 20195October 6, 2018152.782.9/141.819
2169JB DiegoJune 20, 20195October 6, 2018152.788.7/136.220
2270CostelloJune 26, 20193October 7, 2018156.9105.0/145.626
2371JB DiegoJune 27, 20197October 14, 2018150.0110.1/161.723
2475JB DiegoJune 21, 20204October 14, 2019167.378.3/118.015
2576CostelloJuly 2, 20196October 10, 2018156.9101.2/145.322
2679JB DiegoJuly 2, 20204October 14, 2019167.399.2/151.323
2784JB DiegoJuly 5, 20195October 6, 2018152.7107.9/152.025
2885JB DiegoJuly 12, 20198October 12, 2018164.790.2/146.422
2988JB DiegoJuly 16, 20195October 6, 2018152.7110.7/156.026
3092JB DiegoJuly 26, 20204October 14, 2019167.380.6/123.117
3193CostelloJuly 26, 20196October 14, 2018156.988.1/140.319
3294JB DiegoAugust 2, 20195October 6, 2018152.798.6/142.326
3395JB DiegoAugust 13, 20198October 12, 2018164.7100.0/149.525
3496CostelloAugust 13, 20193October 7, 2018156.9102.5/138.724

3.3.

Barley Dataset

The barley seen field-day dataset includes 20 GS classes where each class includes 2000 images for training, 600 images for validation, and 1400 images for test. These 20 classes include four classes in the seedling stage, four classes in the tillering stage, two in stem elongation, one in booting, two in ear emergence, one in milk development, three in the dough development, and three in ripening. There are 80,000 barley images overall in this dataset. The barley training, validation, and test images are from four fields and include two different cultivars of Cassia and Infinity. The brightness range in the barley training data varies between 76.4 and 159.6 (AV). There are four different seed rates in the barley training data. The wind speed at the time of capturing the barley training data varied between 9 and 27  km/h. Figure 2 shows sample barley images from these two cultivars at various GSs.

Fig. 2

Samples of barley image from two cultivars of Infinity and Cassia. Image samples are from downward set of data.

JEI_32_3_033014_f002.png

The unseen field-day barley dataset, which is separate from the seen field-day data. It includes 13 GS classes with 2000 images per class. These classes include two from seedling, four from tillering, two from stem elongation, two from ear emergence, one from dough development, and two from ripening. Each class of test data includes 1000 downward and 1000 images of 45-deg-angled looking. Overall, 26,000 images of unseen field-day barley data are in the test dataset, which was collected from three distinct fields in Ireland. These images include brightness variation from 81.2 to 157.8 (AV) and three different seed rates. The wind speed at the time of capturing barley test data varied between 14 and 23  km/h.

The unseen field-day data are only used for testing and not training. Details of the barley dataset are provided in Table 3 and the seen/unseen field-day split is listed in Table 4(b). Information about the fields and their GPS can be found in Table 5.

Table 3

Barley training and test dataset.

Barley
IndexGrowth stageCultivarDate capturedFieldDate sowed onSeed-rate (kg/ha)Min/max brightness (AV)Wind (km/h)
111InfinityNovember 10, 20199October 7, 2019220.388.0/134.827
213InfinityNovember 1, 201813October 3, 2018172.681.2/156.718
314InfinityNovember 1, 201810September 20, 2018227.5104.6/134.518
415CassiaNovember 30, 201814September 22, 2018197.389.6/123.026
518InfinityNovember 28, 201911September 27, 2019210.7101.5/145.723
619InfinityDecember 10, 201810September 20, 2018227.596.4/139.729
722CassiaFebruary 22, 201914September 22, 2018197.395.2/140.521
822InfinityFebruary 20, 20209October 7, 2019220.383.3/138.319
923InfinityFebruary 20, 201913October 3, 2018172.699.6/148.222
1024InfinityFebruary 21, 201910September 20, 2018227.5100.4/143.520
1124InfinityFebruary 18, 201915October 7, 2018186.385.2/141.623
1227InfinityApril 20, 201913October 3, 2018172.697.1/129.016
1328InfinityMarch 1, 202011September 27, 2019210.7108.5/145.226
1429InfinityMarch 27, 201910September 20, 2018227.5104.2/159.69
1532InfinityMay 1, 201910September 20, 2018227.576.4/115.925
1632InfinityMay 7, 201913October 3, 2018172.6110.3/154.314
1733InfinityMay 1, 201915October 7, 2018186.3108.3/157.817
1834CassiaMay 3, 201912September 16, 2018196.1102.0/153.318
1943InfinityMay 12, 202011September 27, 2019210.793.7/148.222
2052CassiaMay 24, 201912September 16, 2018196.187.2/126.126
2155InfinityJune 12, 201913October 3, 2018172.695.6/139.216
2258InfinityJune 10, 201915October 7, 2018186.3105.0/150.623
2359InfinityJune 10, 201910September 20, 2018227.598.5/146.322
2478CassiaJune 20, 201912September 16, 2018196.198.6/156.619
2580InfinityJune 21, 201910September 20, 2018227.590.0/148.623
2682InfinityJune 21, 20209October 7, 2019220.393.5/146.021
2783InfinityJuly 2, 201913October 3, 2018172.682.6/151.424
2884CassiaJune 27, 201912September 16, 2018196.183.0/136.322
2990CassiaJuly 5, 201912September 16, 2018196.198.6/152.626
3091CassiaJuly 16, 201914September 22, 2018197.3103.4/150.023
3192InfinityJuly 12, 201910September 20, 2018227.5102.0/150.620
3293InfinityJuly 12, 201915October 7, 2018186.390.6/134.217
3395InfinityJuly 26, 201910September 20, 2018227.586.2/134.424

Table 4

Seen and unseen field-day test split by GS (a) wheat dataset and (b) barley dataset.

IndexGrowth stageSeen field-dayUnseen field-day
(a) Wheat
111y
212y
314y
415y
516y
617y
722y
823y
923y
1026y
1126y
1228y
1329y
1430y
1532y
1634y
1752y
1855y
1957y
2059y
2169y
2270y
2371y
2475y
2576y
2679y
2784y
2885y
2988y
3092y
3193y
3294y
3395y
3496y
(b) Barley
111y
213y
314y
415y
518y
619y
722y
822y
923y
1024y
1124y
1227y
1328y
1429y
1532y
1632y
1733y
1834y
1943y
2052y
2155y
2258y
2359y
2478y
2580y
2682y
2783y
2884y
2990y
3091y
3192y
3293y
3395y

Table 5

Fields and their GPS coordinates.

Field numberGPSCountyCrop
152.899117, −6.885672KildareWheat
253.196834, −6.822663KildareWheat
353.206064, −6.842477KildareWheat
453.854804, −6.463052LouthWheat
553.840108, −6.550665LouthWheat
652.899833, −6.880663KildareWheat
752.900737, −6.849316KildareWheat
852.878428, −6.853076KildareWheat
953.194527, −6.821440KildareBarley
1053.208336, −6.845517KildareBarley
1153.855659, −6.445509LouthBarley
1253.837405, −6.550965LouthBarley
1352.906955, −6.882233KildareBarley
1453.306339, −6.528849KildareBarley
1552.870509, −6.842997KildareBarley

4.

Methods and Evaluation

In this section the methods for GS estimation are described and the achieved results are presented. Section 4.1 presents the conventional machine learning algorithm with a SVM classifier. The ConvNet with learning from scratch approach and the ConvNet with transfer learning are presented in Secs. 4.2 and 4.3, respectively.

4.1.

SVM Classifier

GS classification of wheat and barley crops was investigated using feature extraction and an SVM classifier.21 Blurry images were detected using a Laplacian Kernel and were removed from dataset below a threshold of 120.22 Data were pre-processed by brightness correction.23

The best results using the SVM classifier were obtained by training on a mix of downward and 45-deg-angled looking images in each class of data. Excess Green index features24 were extracted from images. Data dimensionality was reduced by employing principal component analysis. The SVM classifier was equipped with a radial basis function kernel,25 and regularization parameters of C=1.0 and γ=0.1. A five-fold cross validation scheme was applied and 1400 images per class were utilized for testing purposes. Moreover, for each crop (wheat/barley) 13 classes of unseen field-day test data were used for principal GS classification.

GS classification using the SVM classifier with input pre-processing and a mix of downward and 45-deg-angled looking images in each class, resulted in 63.8% and 59.8% accuracy for wheat and barley, respectively. Principal GS classification using the same classifier on unseen field-day test data resulted in 26.4% and 29.3% accuracy rates for wheat and barley, respectively. Table 6 presents a summary of the experimental results obtained using the SVM classifier.

Table 6

The result of GS classification and principal GS classification of unseen field-day data using SVM classifier.

CropDownward/45-deg-angled imagesData pre-processingGS classification accuracy (%)Principal GS classification accuracy (%)
WheatDownwardNo42.126.4
BarleyDownwardNo40.629.3
WheatDownwardYes49.126.4
BarleyDownwardYes51.229.3
Wheat45 deg angledYes52.926.4
Barley45 deg angledYes50.729.3
WheatDownward and 45 deg angledNo59.626.4
BarleyDownward and 45 deg angledNo56.229.3
WheatDownward and 45 deg angledYes63.826.4
BarleyDownward and 45 deg angledYes59.829.3

4.2.

ConvNet with Learning from Scratch

Two ConvNet models were trained from scratch for GS image classification and principal GS classification of wheat and barley crops.

The first ConvNet includes five-trainable layers including three Conv layers and two dense layers. The Conv layers (Conv2D, Conv2D-1, Conv2D-2) have 32, 64, and 64 filters respectively and the filter size was set to 3×3 and the dense-layers have 1024 and 21/20 neurons for wheat/barley, respectively.

The second ConvNet is almost identical to the first ConvNet apart from two layers of batch normalization that are added to the network after the max-pooling layer of the first and the third trainable layers. The following paragraphs are summarized and the equations were removed.

For all ConvNet experiments, an image size of 256×256 was employed. An image size of 125×125 was tested but the results were not satisfactory. The image pixel values were rescaled to [0,1] interval. The training data were pre-processed using the hue, saturation, value color space, employing the brightness correction function.

Since data augmentation has proven effective in training deep learning algorithms.26 Three different data augmentation schemes were applied to the input of the network. The data were augmented with various in-range brightness values.27 To this end, while reading the images into the training data-generator, the brightness range was set to produce either darker images (setting uniform distribution values <1.0) or brighter images (by setting uniform distribution values over 1.0). See Eq. (1) for the transform equation and Table 7 for the brightness parameter setting. In this work, the brightness range is set to [0.7,1.3] for data brightness augmentation

Eq. (1)

x=x+δ.

Table 7

Description and range of values of the parameters used for brightness, rotation, and zoom augmentation, where u(a,b) denotes a uniform distribution.

ParameterDescriptionRange
δBrightnessu(0.7,1.3)
θRotation angleu(90  deg,90  deg)
zxHorizontal scaleu(0.7,1.3)
zyVertical scaleu(0.7,1.3)

The network was made robust,28 to 90-deg rotation ranges by data rotation augmentation.27 This method randomly rotates the image clockwise by the given angle. The affine transformation for rotation can be found in Eq. (2). The rotation range parameter setting employed in this work is listed in Table 7.

Eq. (2)

[xy]=[zxcos(θ)zysin(θ)zxsin(θ)zycos(θ)][xy].

Zoom augmentation27 was exerted on the training data by employing a scale range of [0.7,1.3]. The affine transform for zoom augmentation complies with Eq. (2) and the parameter settings of horizontal and vertical scale were obtained from Table 7. This function randomly produces images that are zoomed in for values <1.0 (interpolates original image pixel values) and zoomed out for values greater than 1.0 (add new pixel values around the original image).

The input for each class of data includes 50% downward and 50% angled images. The test data were classified in the principal GS range without any pre-processing.

An extensive series of experiments were carried out to find the best performing ConvNet with learning from scratch model and input format for the GS classification and principal GS classification task. The results demonstrate an improvement when employing the ConvNet with batch normalization layers. Moreover, including input pre-processing and data augmentation was proven to play an important role for both the classification and principal GS classification tasks using this network, see Table 8.

Table 8

Overall performance of the ConvNet trained from scratch for GS classification on test data and principal GS classification on unseen field-day test data for wheat and barley crops. There are three different experiments: whether (a) the network includes batch normalization layers, (b) the input data is a mix of downward and 45-deg-angled looking images, and (c) training includes pre-processing and data augmentation

Crop(a)(b)(c)GS classification accuracy (%)Principal GS classification accuracy (%)
WheatNoNoNo88.642.3
BarleyNoNoNo88.241.4
WheatYesNoNo90.349.1
BarleyYesNoNo91.152.3
WheatNoYesYes92.170.3
BarleyNoYesYes92.768.3
WheatYesYesYes95.473.4
BarleyYesYesYes96.577.2

The best average results for barley and wheat using the ConvNet learned from scratch including batch normalization, input pre-processing, and data augmentation is 95.9% GS classification accuracy, and principal GS classification accuracy of 75.3%.

4.3.

ConvNet with Transfer Learning

The ConvNet with transfer learning approach seeks to transfer knowledge from a source task to a target task. The network’s pre-trained parameters from the source task are re-purposed for a target task within a similar or related domain. The concept of transfer learning relaxes the urgent requirement of having the ConvNet trained on a large independent and identically distributed dataset.

The following paragraphs are summarized and the description of experiments are removed.

The Visual Geometry Group ConvNets has provided a reliable base for numerous image recognition systems since its introduction in 2014.29 In this work, Visual Geometry Group-19 was employed as a basis for transfer learning, with 19 weight layers of 16 Conv and three fully connected (FC) layers. The network is pre-trained on the ImageNet dataset and the knowledge can be transferred at any layer of the network for the new classification task. To acquire the best architecture for setting trainable and non-trainable layers for the classification problem at hand, five different experiments were tested. The experiment models of E1 to E5 are listed in Table 9, including non-trainable and trainable layers and parameters.

Table 9

Series of experiments to find the best-performing transfer learning method, the table includes the description and number of trainable parameters.

EXPModel layers21 classes of wheat20 classes of barley
Non-trainable layersTrainable layersNo. of total parametersNo. of trainable parametersNo. of total parametersNo. of trainable parameters
E116 Conv layersTwo FC layers20,571,221546,83720,570,196545,812
E216 Conv layersThree FC layers21,085,2691,060,88521,084,7561,060,372
E316 Conv layersFour FC layers21,211,2211,186,83721,210,9641,186,580
E415 Conv layersOne Conv layer and21,211,2213,546,64521,210,9643,546,388
Four FC layers
E514 Conv layersTwo Conv layers and21,211,2215,906,45321,210,9645,906,196
Four FC layers

The input for each class of data includes 50% downward and 50% angled images. Similar to the data preparation of ConvNet with learning from scratch (presented in Sec. 4.2), pre-processing and data augmentation were applied to the training data of this network. The training data are augmented for brightness27 in range [0.7,1.3], rotation in the range 90 deg28 and zoom augmentation with scale in the range [0.7,1.3].27 The test data were classified in the principal GS range without any pre-processing or data augmentation.

Among the aforementioned experiments, Experiment E4 achieved the best GS classification accuracy, with 15 non-trainable Conv-layers, including the last Conv-layer and four FC-layers of 1024, 512, 256, 21/20 nodes as trainable layers. The input setting for this practice includes data pre-processing, data augmentation and a mix of downward and 45-deg-angled images in each class of data.

The result from Experiments E1 to E5 were investigated to choose the best transfer learning model, Fig. 3. The accuracies for experiment E1 are 76.2% and 73.4% for wheat and barley respectively, these increased to 99.1% and 99.7% in experiment E4. Moving deeper by training the network with another trainable ConvNet layer, accuracy drops slightly to 98.1% and 98.3% for wheat and barley, respectively. This is a costly drop in accuracy as the number of trainable parameters doubles from 3.5 to 6 million.

Fig. 3

GS classification accuracy rate, including network’s number of trainable parameters for each experiment. Model trained on (a) wheat and (b) barley data.

JEI_32_3_033014_f003.png

It is shown that experiment E4 resulted in the best performance with a reasonable number of trainable parameters. Hence it was used as the base model for training and testing the GS classification.

The results obtained for the various methods of GS classification and principal GS classification employing experiment E4 are presented in Table 10.

Table 10

Overall performance of ConvNet with transfer learning experiment E4, for GS classification and principal GS classification of wheat and barley. There are three different experiments: whether (a) the input data is a mix of downward and 45-deg-angled looking images, (b) training data includes pre-processing, and (c) training data includes data augmentation.

Crop(a)(b)(c)GS classification accuracy (%)Principal GS classification accuracy (%)
WheatNoNoNo93.673.1
BarleyNoNoNo92.470.4
WheatNoYesNo93.777.6
BarleyNoYesNo92.175.3
WheatNoYesYes95.383.8
BarleyNoYesYes93.185.2
WheatYesYesYes97.393.5
BarleyYesYesYes97.592.2

The first two rows of Table 10 present the results of training with single mode data (either angled or downward looking images) with no pre-processing or data augmentation. The result of GS classification using these inputs is fairly good with 93.6% and 92.4% accuracy for wheat and barley, respectively. However, principal GS classification using the aforementioned trained network does not yield good results for unseen field-day data; generating principal GS classification accuracy of 73.1% and 70.4% for wheat and barley, respectively.

The next series of experiments involved training the transfer learning network using pre-processed data with brightness correction. The network classifies GSs with almost the same accuracy rates as the previous experiment. However principal GS classification improved noticeably reaching 77.6% and 75.3% accuracy for wheat and barley, respectively.

Including input data augmentation as well as pre-processing brings about higher GS classification and principal GS classification accuracies. The results show 95.3% and 93.1% accuracy for wheat and barley GS classification. Moreover the principal GS classification progressed positively and achieved 83.8% and 85.2% accuracy for wheat and barley, respectively.

Finally, including downward and 45-deg-angled images in each class of data for training was considered. The input data were pre-processed and augmented as well. A significant improvement was noticed in both classification accuracy rates and the number of images from each class correctly classified in their corresponding principal GSs. Employing this input setting to transfer learning, the network achieved 97.3% and 97.5% GS classification accuracy rates and principal GS classification accuracies of 93.5% and 92.2% were achieved for wheat and barley, respectively, see Table 10.

For both GS classification and principal GS classification tasks, the network architecture of experiment E4 trained on a mix of downward and 45-deg-angled looking images in each class, including pre-processing and data augmentation, achieved the best results. The confusion matrices for principal GS classification of wheat and barley unseen field-days data are presented in Figs. 4(a) and 4(b), respectively.

Fig. 4

(a) Wheat and (b) barley, principal GS classification employing ConvNet with transfer learning experiment E4, presented in confusion matrix of actual GSs on left and the classified principal GS range at top.

JEI_32_3_033014_f004.png

5.

Discussion

Of the three methods considered, the ConvNet with transfer learning including data pre-processing and a mix of downward and 45-deg-angled looking for training, resulted in the best GS classification accuracy for both wheat and barley crops. As shown in Fig. 5, image pre-processing together with a mix of downward and 45-deg-angled looking images for training, produced the best classification accuracy for each method.

Fig. 5

Comparison of the GS classification accuracy for three different methods of (I) feature extraction and SVM classifier, (II) ConvNet with learning from scratch, and (III) ConvNet with transfer learning experiment E4, employed on wheat and barley data. The accuracy achieved for each method averaged over test downward and 45-deg-angled images and is reported in different practices of (a), (b), and (c) as follows: (a) experiment with no data pre-processing, (b) experiment with data pre-processing, and (c) experiment with data pre-processing and mix of downward and 45-deg-angled looking input images.

JEI_32_3_033014_f005.png

The evaluation of principal GS classification shows that of the three methods, the ConvNet with transfer learning achieved the highest accuracy. The principal GS classification accuracies achieved for wheat and barley crops using this method was 93.5% and 92.2%, respectively. As shown in Fig. 6, image pre-processing together with a mix of downward and 45-deg-angled looking images as the input for training produced the best principal GS classification for each method.

Fig. 6

Comparison of the principal GS classification accuracy for three different methods of (I) feature extraction and SVM classifier, (II) ConvNet with learning from scratch, and (III) ConvNet with transfer learning experiment E4, employed on wheat and barley unseen field-day data. The accuracy achieved for each method averaged over test downward and 45-deg-angled images and is reported in different practices of (a), (b), and (c) as follows: (a) experiment with no data pre-processing, (b) experiment with data pre-processing, and (c) experiment with data pre-processing and mix of downward and 45-deg-angled looking images.

JEI_32_3_033014_f006.png

The result of principal GS classification for 13-classes (26,000 images of unseen field-day) of wheat is presented in the confusion matrix, Fig. 4(a). The accuracy achieved for wheat principal GS classification was 93.5%.

Likewise, the result of principal GS classification for 13-classes (26,000 images of unseen field-day) of barley is presented in the confusion matrix in Fig. 4(b). The accuracy achieved for barley principal GS classification was 92.2%.

6.

Conclusion

Evaluation of three different machine learning methods along with three methods of crop GS classification of wheat and barley are presented in this paper.

Of the three methods, the ConvNet with transfer learning including data pre-processing and a mix of downward and 45-deg-angled looking for training, resulted in the best GS classification accuracy for both wheat and barley crops. As shown in Fig. 5, image pre-processing together with a mix of downward and 45-deg-angled looking images for training, produced the best classification accuracy for each method.

Moreover, the evaluation of principal GS classification showed that of the three methods, the ConvNet with transfer learning achieved the highest accuracy. The principal GS classification accuracy achieved for wheat and barley crops using this method was 93.5% and 92.2%, respectively. As shown in Fig. 6, image pre-processing together with a mix of downward and 45-deg-angled looking images as the input for training produced the best principal GS classification for each method.

The results of classification of downward and 45-deg-angled images show that the ConvNets yielded higher classification accuracy for angled looking images, see Fig. 7(a). Principal GS classification also demonstrated a similar trend of yielding better principal GS classification accuracy for angled looking images while employing ConvNet models, see Fig. 7(b).

Fig. 7

Employing best practices of method (I) feature extraction and SVM classifier, (II) ConvNet with learning from scratch, and (III) ConvNet with transfer learning experiment E4 for (a) GS classification of downward and 45-deg-angled looking images. (b) Principal GS classification of unseen field-day test data of downward and 45-deg-angled looking images.

JEI_32_3_033014_f007.png

Detailed evaluation of principal GS classification results for various GSs of wheat shows that except for downward images of GS32, unseen field-day data were classified in their principal GSs with acceptable accuracy. Likewise, evaluation of principal GS classification results for various GSs of barley shows that downward images of GS32 has the worst of the principal GS classification results. GS32 is the stage in which the canopy closure happens and leaf area index (LAI) reaches its saturation point.

The key novel contributions of this work include development of a unique labeled dataset of proximal images of wheat and barley crop GSs, GS classification of cereal crops using ConvNet with transfer learning, ConvNet with learning from scratch, and SVM classifier. Moreover, this research is the first comparison of these methods for the problem of cereal GS classification.

In future work, the existing image dataset could be augmented by employing state-of-the art image synthesis algorithms, such as texture synthesis, image super resolution,30 and generative adversarial networks.31 Although our existing dataset is large enough for training neural networks, it only includes two variety of the crops. Perhaps a larger dataset including more crop varieties could further improve the principal GS classification.

The comparison of results with different data types to determine the best performing model shows that including images with two different camera views boosts the performance of principal GS classification dramatically. Hence, a more robust trained network may be obtained using training images from several camera view angles.

An unsupervised learning algorithm, such as an unsupervised deep learning algorithm,32 could be used for GS classification. The trained network for GS classification of wheat and barley crops may be applicable to GS classification of other cereal crops with similar visual GSs, such as rye, triticale, and oats.

Data Availablity

The data that supports the findings of this study are available from “CONSUS Program and Origin Enterprises Plc” repository but restrictions apply to the availability of these data, which were used under license number (16/SPP/3296) for the current study and so are not publicly available. Data are, however, available from the authors upon reasonable request and with permission from “CONSUS Program and Origin Enterprises Plc” authorities. If you require any further information, please do not hesitate to contact the author by email.

Acknowledgments

This research forms part of the CONSUS program which is funded under the Science Foundation Ireland Strategic Partnerships Program (Grant No. 16/SPP/3296) and is co-funded by Origin Enterprises Plc. The authors would like to thank Lyons Research Farm of UCD, Irish farmers in County Louth: J., P. & T. McGuiness, B. Lynch and P. O’Grady, K. Dowling in Kildare, for their kind co-operation during the data collection and also providing us with the metadata for their fields. The authors declare that they have no conflict of interest regarding the publication of this paper.

References

1. 

D. Tilman et al., “Global food demand and the sustainable intensification of agriculture,” Proc. Natl. Acad. Sci. U. S. A., 108 (50), 20260 –20264 https://doi.org/10.1073/pnas.1116437108 (2011). Google Scholar

2. 

D. Feng et al., “Advances in plant nutrition diagnosis based on remote sensing and computer application,” Neural Comput. Appl., 32 1 –10 https://doi.org/10.1007/s00521-018-3932-0 (2019). Google Scholar

3. 

V. Wiley and T. Lucas, “Computer vision and image processing: a paper review,” Int. J. Artif. Intell. Res., 2 (1), 22 –36 https://doi.org/10.29099/ijair.v2i1.42 (2018). Google Scholar

4. 

S. Rasti et al., “A survey of high resolution image processing techniques for cereal crop growth monitoring,” Inf. Process. Agric., 9 300 –315 https://doi.org/10.1016/j.inpa.2021.02.005 (2021). Google Scholar

5. 

W. Thomas, “The value of decimal cereal growth stages,” Ann. Appl. Biol., 165 (3), 303 –304 https://doi.org/10.1111/aab.12145 AABIAV 0003-4746 (2014). Google Scholar

6. 

Agriculture and H. D. B. 2018, “Wheat growth guide,” 1–42 (2018). https://horticulture.ahdb.org.uk/ Google Scholar

7. 

J. C. Zadoks, T. T. Chang and C. F. Konzak, “A decimal code for the growth stages of cereals,” Weed Res., 14 (6), 415 –421 https://doi.org/10.1111/j.1365-3180.1974.tb01084.x WEREAT 1365-3180 (1974). Google Scholar

8. 

H. Bleiholder et al., Growth Stages of Mono-and Dicotyledonous Plants, 158 Federal Biological Research Centre for Agriculture and Forestry, Berlin/Braunschweig (2001). Google Scholar

9. 

V. Subramanian, T. F. Burks and A. Arroyo, “Development of machine vision and laser radar based autonomous vehicle guidance systems for citrus grove navigation,” Comput. Electron. Agric., 53 (2), 130 –143 https://doi.org/10.1016/j.compag.2006.06.001 CEAGE6 0168-1699 (2006). Google Scholar

10. 

E. R. Hunt et al., “Acquisition of nir-green-blue digital photographs from unmanned aircraft for crop monitoring,” Remote Sens., 2 (1), 290 –305 https://doi.org/10.3390/rs2010290 RSEND3 (2010). Google Scholar

11. 

J. Xue, L. Zhang and T. E. Grift, “Variable field-of-view machine vision based row guidance of an agricultural robot,” Comput. Electron. Agric., 84 85 –91 https://doi.org/10.1016/j.compag.2012.02.009 CEAGE6 0168-1699 (2012). Google Scholar

12. 

Z. Yu et al., “Automatic image-based detection technology for two critical growth stages of maize: emergence and three-leaf stage,” Agric. For. Meteorol., 174 65 –84 https://doi.org/10.1016/j.agrformet.2013.02.011 0168-1923 (2013). Google Scholar

13. 

P. Sadeghi-Tehran et al., “Automated method to determine two critical growth stages of wheat: heading and flowering,” Front Plant Sci, 8 252 –266 https://doi.org/10.3389/fpls.2017.00252 (2017). Google Scholar

14. 

S. Zhao et al., “Rapid yield prediction in paddy fields based on 2D image modelling of rice panicles,” Comput. Electron. Agric., 162 759 –766 https://doi.org/10.1016/j.compag.2019.05.020 CEAGE6 0168-1699 (2019). Google Scholar

15. 

Z. Zainuddin et al., “Rice farming age detection use drone based on SVM histogram image classification,” J. Phys. Conf. Ser., 1198 (9), 092001 https://doi.org/10.1088/1742-6596/1198/9/092001 JPCSDZ 1742-6588 (2019). Google Scholar

16. 

A. Yudhana, R. Umar and F. M. Ayudewi, “The monitoring of corn sprouts growth using the region growing methods,” J. Phys. Conf. Ser., 1373 (1), 012054 https://doi.org/10.1088/1742-6596/1373/1/012054 JPCSDZ 1742-6588 (2019). Google Scholar

17. 

S. Rasti et al., “Crop growth stage estimation prior to canopy closure using deep learning algorithms,” Neural Comput. Appl., 33 (5), 1733 –1743 https://doi.org/10.1007/s00521-020-05064-6 (2021). Google Scholar

18. 

I. Goodfellow et al., Deep Learning, 1 MIT Press, Cambridge (2016). Google Scholar

19. 

A. Krizhevsky, I. Sutskever and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Adv. Neural Inf. Process. Syst., 25 1097 –1105 https://doi.org/10.1145/3065386 1049-5258 (2012). Google Scholar

20. 

“DJI Official Website,” https://www.dji.com/ (2018). Google Scholar

21. 

L. Wang, Support Vector Machines: Theory and Applications, 177 Springer Science & Business Media( (2005). Google Scholar

22. 

Z. Al-Ameen et al., “A comprehensive study on fast image deblurring techniques,” Int. J. Adv. Sci. Technol., 44 (2012). Google Scholar

23. 

Y. Hatanaka et al., “Improvement of automatic hemorrhage detection methods using brightness correction on fundus images,” Proc. SPIE, 6915 69153E https://doi.org/10.1117/12.771051 PSISDG 0277-786X (2008). Google Scholar

24. 

G. E. Meyer, T. W. Hindman and K. Laksmi, “Machine vision detection parameters for plant species identification,” Proc. SPIE, 3543 327 –336 https://doi.org/10.1117/12.336896 (1999). Google Scholar

25. 

C. J. Burges, “A tutorial on support vector machines for pattern recognition,” Data Mining Knowl. Discov., 2 (2), 121 –167 https://doi.org/10.1023/A:1009715923555 (1998). Google Scholar

26. 

L. Perez and J. Wang, “The effectiveness of data augmentation in image classification using deep learning,” (2017). Google Scholar

27. 

A. Hernandez-Garcia, “Data augmentation and image understanding,” (2020). Google Scholar

28. 

C. Shorten and T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” J. Big Data, 6 (1), 1 –48 https://doi.org/10.1186/s40537-019-0197-0 (2019). Google Scholar

29. 

K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” (2014). Google Scholar

30. 

X. Wu, K. Xu and P. Hall, “A survey of image synthesis and editing with generative adversarial networks,” Tsinghua Sci. Technol., 22 (6), 660 –674 https://doi.org/10.23919/TST.2017.8195348 (2017). Google Scholar

31. 

A. Creswell et al., “Generative adversarial networks: an overview,” IEEE Signal Process Mag., 35 (1), 53 –65 https://doi.org/10.1109/MSP.2017.2765202 ISPRE6 1053-5888 (2018). Google Scholar

32. 

J. Huang et al., “Unsupervised deep learning by neighbourhood discovery,” in Int. Conf. Mach. Learn., 2849 –2858 (2019). Google Scholar

Biography

Sanaz Rasti is a research software engineer in the School of Computer Science, University College Dublin, Ireland. She is a highly motivated researcher in the field of machine learning (ML) and artificial intelligence. She has hands-on experience in a wide range of ML applications including computer vision and natural language processing. She has been awarded for her research by Science Foundation Ireland, Origin Enterprises plc, and Irish Research Council.

Chris J. Bleakley is an associate professor and head of the School of Computer Science. His research focuses on pervasive computing, in particular indoor positioning systems and networked embedded systems. His work has applications in indoor navigation, smart agriculture, biomedical devices, and environmental monitoring. To date, he has been awarded over €2 million in external research funding as principle investigator. He has graduated 12 PhD students and 4 MSc by research students as sole supervisor. To date, he has published 1 book, 42 journal papers, and 68 conference papers.

Gregory M. P. O’Hare is a professor of artificial intelligence and head of the School of Computer Science and Statistics at Trinity College Dublin. He has published over 500 refereed publications in journals and international conferences, 7 books and has won significant grant income (CA €48.00M). He is an established researcher of international repute. His research interests are in the areas of distributed artificial intelligence and multi-agent systems, intelligent systems, ubiquitous computing, and wireless sensor networks.

Biographies of the other authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Sanaz Rasti, Chris J. Bleakley, Guénolé C. M. Silvestre, Gregory M. P. O’Hare, and David Langton "Assessment of deep learning methods for classification of cereal crop growth stage pre and post canopy closure," Journal of Electronic Imaging 32(3), 033014 (31 May 2023). https://doi.org/10.1117/1.JEI.32.3.033014
Received: 7 October 2022; Accepted: 30 April 2023; Published: 31 May 2023
Advertisement
Advertisement
KEYWORDS
Education and training

Image classification

Deep learning

Machine learning

Data modeling

Ear

Cameras

Back to Top