In the research process, the classification effect of the U-Net deep semantic segmentation network on the planting structure is found to be better than those of the FCN network and the SegNet network model. Therefore, in this section, only U-Net and its improved models are compared and analyzed. In this paper, the conventional U-Net model, the U-Net model with an added deep, separable convolution module, and the activation function network model are trained in 200 batches on the same test dataset to obtain the accuracy curves shown in the following figure 4.
Figure 4. Results of U-Net and its improved models obtained from a comparison diagram: (a) U-Net training accuracy rate curve; (b) Separable-UNet training accuracy rate curve; (c) Mish-Separable-UNet training accuracy rate curve.
In this paper, the G-S image fusion method was used in combination with a deep semantic segmentation network to provide valuable spectral feature information. That is, this method was applied to improve the classification accuracy of the planting structure.
Compared with the traditional U-Net model, after adding the depth-separable convolution module, the model reaches an overfitting batch from 100 to 30, significantly reducing the calculation cost and increasing the fitting rate. After using the Mish activation function, the accuracy and calculation rate are slightly increased due to the smoothing characteristics of the function. The improved network model achieves lighter parameters, and the weight file size is reduced from
303 MB to 53 MB, a reduction of 82.5%, significantly improving the production cost.
The original remote sensing images of seven bands from 1 to 7 are synthesized with 1, 2, and 3 exponents. Then, the G-S method is used to fuse the panchromatic groups to further increase the spatial resolution and spectral richness of the training images and predict the results, as shown in the picture below.
The statistical test set prediction results are compared with the label data to generate the confusion matrices shown in Figure 6 below to judge the classification quality of the planting structure.
Figure 6. The classification result from confusion matrix (a) seven-band confusion matrix; (b) synthetic NDVI confusion matrix; (c) synthetic AWEI confusion matrix; (d) synthetic SAVI confusion matrix; (e) synthetic NDVI and AWEI confusion matrix; (f) Synthesis of NDVI and SAVI confusion matrix; (g) Synthesis of AWEI and SAVI confusion matrix; (h) Synthesis of three exponential confusion matrices.
Table 1 compares the prediction results of the various methods with the actual label data to calculate the accuracy rate. After the investigation, most of the woodland area is found to comprise economic crops such as jujube trees, so wheat, woodland, and cotton areas are classified as crop areas.
TABLE 1
Classification accuracy statistical table.
Index
|
Water
|
Wheat
|
Forest
|
Cities
|
Cotton
|
Crops
|
Accuracy
|
No Indices
|
80.6%
|
77.5%
|
71.0%
|
95.0%
|
86.9%
|
81.7%
|
90.3%
|
NDVI
|
81.4%
|
84.3%
|
73.8%
|
92.8%
|
86.6%
|
83.5%
|
92.4%
|
AWEI
|
81.4%
|
83.9%
|
74.2%
|
93.9%
|
88.2%
|
84.1%
|
92.7%
|
SAVI
|
80.3%
|
84.4%
|
69.7%
|
93.3%
|
85.6%
|
82.4%
|
91.6%
|
NDVI, AWEI
|
81.8%
|
85.2%
|
74.4%
|
93.5%
|
87.5%
|
84.2%
|
93.7%
|
NDVI, SAVI
|
81.0%
|
82.3%
|
73.8%
|
93.1%
|
87.2%
|
83.4%
|
92.6%
|
AWEI, SAVI
|
82.1%
|
81.8%
|
74.4%
|
93.5%
|
87.5%
|
83.9%
|
93.3%
|
Three Indices
|
81.5%
|
85.3%
|
74.1%
|
93.0%
|
87.8%
|
84.1%
|
93.8%
|
As shown in Table 1, the overall accuracy of the 7-band fusion image without the addition of any index reached 90.3%, and the accuracy of increasing the index image participation in the image fusion process was higher than that obtained without the index. Adding the index uses the improved U-Net network in the training process for identifying crop cultivation areas, the structural classification has a positive effect on the results.
With the addition of each of the three indexes individually, all showed an increase in the overall accuracy rate, among which AWEI had the most significant impact on the classification of the planting structure. The crop identification accuracy was increased by 2.4%, and the city and water pixels were better identified. This index is superior to the other two indexes, which reached accuracies of 81.4% and 93.9%, and the improvement effect is visible. The overall accuracy rate reaches 92.7%, which is the best among the three indexes. However, this index tends to judge pixels as cotton when recognizing mixed pixels of wheat and cotton, and its ability to identify crops in a large mixed area remains to be discussed.
Under the condition of two-index 9-band fusion images, AWEI still plays an important role. The classification effect of training images containing AWEI is better than that of the NDVI-SAVI 9-band fusion images. The addition of NDVI further improves the crop recognition ability of AWEI and the mutual recognition ability among crops. The accuracy rates of the three types of plants, wheat, cotton, and woodland, increased to 85.2%, 74.4%, and 87.5%, respectively. At the same time, this method also guarantees that the high recognition accuracy of water and cities reaches 81.8% and 93.5%, respectively. Only the water identification efficiency is lower than that of the AWEI-SAVI method, with a difference of 0.3%, and the overall accuracy rate of this method reaches 93.7%.
Compared with the NDVI-AWEI 9-band fusion image method, the overall accuracy of the 10-band fusion image is improved by only 0.1%, and the crop classification recognition is reduced. SAVI's improvement in the fusion image is limited. Considering the production costs, the priority is lower. In summary, the NDVI-AWEI 9-band fusion image method combined with the improved U-Net network model classification planting structure method has advantages.
This paper evaluates the reliability and accuracy of the planting structure classification results of each index method combined with the improved U-Net model by calculating the kappa coefficient, precision, recall and harmonic mean F1, as shown in Table 2 below. Compared with nonindexed images, the classification accuracy of the indexed fusion image planting structure is significantly improved. The kappa coefficient increases from 0.868 to more than 0.874 with the addition of the index, further improving the consistency between the predicted class and the artificial label. The accuracy, recall, and average value are enhanced by more than 0.02. In the new method, the NDVI- AWEI fusion method has the same accuracy as the three-index method, reaching 0.873. The kappa coefficient and F1 value are better than the three-index method, at 0.886 and 0.872, respectively; thus, this method can provide accurate and stable classification results of the crop planting structure. As shown in Figure 5, the classification confusion in the boundary areas of feature types is still the primary source of error. Figure 7 below shows the statistics of the confusion types and their probabilities under various methods.
TABLE 2
Statistics of each method
Band
|
Index
|
Precision
|
Recall
|
F1
|
Kappa
|
7
|
No Indices
|
0.853
|
0.851
|
0.852
|
0.868
|
8
|
NDVI
|
0.866
|
0.865
|
0.865
|
0.880
|
AWEI
|
0.871
|
0.869
|
0.870
|
0.885
|
SAVI
|
0.861
|
0.856
|
0.858
|
0.874
|
9
|
NDVI, AWEI
|
0.873
|
0.871
|
0.872
|
0.886
|
NDVI, SAVI
|
0.863
|
0.862
|
0.863
|
0.878
|
AWEI, SAVI
|
0.868
|
0.867
|
0.867
|
0.882
|
10
|
Three Indices
|
0.873
|
0.869
|
0.871
|
0.885
|
Due to the increase in spectral richness and the increase in the information contained in the training images, the feature information extracted using the improved U-Net network is enhanced to a certain extent. The overall confusion rate of the prediction results declines, but the edge is not fundamentally resolved. Regarding misclassifications, forest pixels still maintain the highest probability of confusion due to their discrete nature. Using the improved U-Net network to train the NDVI-AWEI 9-band fusion image method compared to other means, the minimum confusion level in each misjudgment type can be maintained at 2%-9%, representing the improved U-Net model planting structure classification model.