Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Multi-feature sparse representation based on adaptive graph constraint for cropland delineation

Open Access Open Access

Abstract

Cropland delineation is the basis of agricultural resource surveys and many algorithms for plot identification have been studied. However, there is still a vacancy in SRC for cropland delineation with the high-dimensional data extracted from UAV RGB photographs. In order to address this problem, a new sparsity-based classification algorithm is proposed. Firstly, the multi-feature association sparse model is designed by extracting the multi-feature of UAV RGB photographs. Next, the samples with similar characteristics are hunted with the breadth-first principle to construct a shape-adaptive window for each test. Finally, an algorithm, multi-feature sparse representation based on adaptive graph constraint (AMFSR), is obtained by solving the optimal objective iteratively. Experimental results show that the overall accuracy (OA) of AMFSR reaches 92.3546% and the Kappa is greater than 0.8. Furthermore, experiments have demonstrated that the model also has a generalization ability.

© 2024 Optica Publishing Group under the terms of the Optica Open Access Publishing Agreement

1. Introduction

Agricultural resources investigations are the basis for obtaining important data in agricultural production [1]. In recent years, UAV photographs have been introduced to agricultural resources investigations because UAV images is convenient and extremely low cost, and it can satisfy large-scale high-precision mapping in local areas [2]. Therefore, it has become an ideal tool for surveying agricultural resources with high precision in local areas. As the application of regional agricultural resources investigations becomes more extensive and there is an ever-growing demand for accuracy as well, the existing algorithms cannot follow the new requirement in high precision. A novel approach should be explored to meet the practical applications of UAV photography.

Cropland is a fundamental resource in agriculture [3]. Accurate recognition and segmentation of cropland in UAV photos is widely recognized in agricultural production and has high application value. Therefore, high-precision recognition and segmentation of cropland in UAV photographs is worth in-depth research.

In the recognition and segmentation of agricultural tasks, previous scholars had done a lot of work [4]. Hyperspectral image (HSI) was introduced to these studies, which involved near-infrared band (NIR) images especially [5]. And they proposed a series of algorithms such as support vector machines(SVM) [6], the random forest [7], network analysis [8], linear representation-based classification(LRC) [9], and so on.

About research images, the references show that HSI seems to be the mainstream of recognition and segmentation in agricultural tasks because the different wavelengths can provide rich spectral characteristics and the near-infrared band (NIR) is sensitive to crops. However, hyperspectral cameras usually require longer exposure times to maintain the signal-to-noise ratio of each spectral channel. And its expensive cost is one of the main constraints that prevent HSI from being widely applied [10]. At the same time, high-precision recognition and segmentation require the image with a higher spatial resolution and a high spatial resolution image is beneficial for accurate manual labeling. However, the higher resolution of the image will magnify the noise that the NIR spectrum brings into it. So, the NIR has a negative impact on high spatial resolution images [11]. Compared to HSI, traditional RGB images have high utility [12]. It is promising research and will significantly improve the efficiency of agricultural resources investigations if the visible UAV photograph can be used to quickly and accurately identify the cropland.

On algorithms based on the visible photograph in agriculture, the references show that they have made impressive progress. To encourage research in computer vision for agriculture, Wang et al. reviewed the applications of machine vision in agriculture [13]. Syazwani et al. present an automatic method for identifying the pineapple’s crown images in the designated plot and they further count the detected images by machine learning classifiers [14]. A leaf area index (LAI) estimation model developed by Liu et al. was compared to other practical methods [15]. They noted that UAVs with RGB cameras are a low-cost and efficient tool. Biswas et al. developed a new method to delineate individual patches and to estimate mangrove cover from high-resolution aerial photography (0.08 m spatial resolution) with RGB channels [16]. Hasan et al. recognize the cropland with winter wheat in Xinjiang, China, via LAI from UAV RGB images [17]. However, in real application scenarios, many data sets are high-dimensional [18]. Classification of high-dimensional data sets cannot be solved merely by traditional methods. Gradually, sparse representation classification (SRC) algorithms entered the scholar’s vision [19]. In SRC, a high-dimensional test can be linearly reconstructed by atoms and its coefficient is sparse. To calculate the residuals between the original test and the reconstructed data in each subspace to classify it into the class in the subspace where its residual is least. While many sparse models have been proposed and several applications have emerged, the study of cropland recognition problems based on UAV RGB photographs is lacking. Thus, we will focus on the SRC study for cropland recognition based on UAV RGB photographs.

The rest of this paper is organized as follows. In section 2, we briefly overview the related work. Section 3 introduces the multi-feature sparse representation based on adaptive graph constraint (AMFSR). In this section, we optimize the objective of AMFSR and then obtain the analytic solution of the model based on strict derivation, which is the key to designing a greedy algorithm. Section 4 presents the experimental results and analyses.

2. Related works

Sparse representation classification(SRC) was proposed by Hui et al. for face recognition [20]. It indicates that the features of samples belonging to the same class are located in the same low-dimensional subspace. And it implies that the most elements in the coefficient vector are zero and SRC can only handle non-zero elements, which greatly reduces the complexity of matrix operations. Because of its excellent ’parse’ property and its high performance of processing high-dimensional data, SRC has been becoming a hotspot in the high-dimensional data classification, such as remote sensing images [21] and biological image recognition [22]. Xun et al. applies the sparse representation algorithm into agricultural aerial imagery classification in 2021 [23].

2.1 Original sparse representation classification

SRC supposes that there are M distinct classes and M dictionaries in the classifying task, in which the dictionary of the m-th class is marked as $D_{m}$ and there are $N_{m}$ atoms in $D_{m}$. If sample $x$ belongs to class m, it can be linearly approximated by the atoms in $D_{m}$.

$$x \approx d_{m}^{1}\alpha_{m}^{1}+d_{m}^{1}\alpha_{m}^{1}+\cdots+d_{m}^{N_{m}}\alpha_{m}^{N_{m}} = \left[d_{m}^{1} \ d_{m}^{2} \ \cdots d_{m}^{N_{m}}\right] \left[\alpha_{m}^{1} \ \alpha_{m}^{2} \ \cdots\ \alpha_{m}^{N_{m}}\right]^{T}$$
where $\alpha _{m}^{n}$ is a coefficient of $d_{m}^{n}$ and $D_{m} = \left [d_{m}^{1} \ d_{m}^{2} \ \cdots \ d_{m}^{N_{m}}\right ]$.

All of $D_{m}$ constitutes the general dictionary $D = [D_{\rm 1} \ D_{\rm 2} \ \cdots \ D_{\rm M}]$, then sample $x$ can still be described as

$$\begin{aligned} x & \approx D_{1}\alpha_{1}+D_{2}\alpha_{2}+\cdots+D_{M}\alpha_{M} = \left[D_{1} \ D_{2} \ \cdots\ D_{M}\right] \left[\alpha_{1} \ \alpha_{2} \ \ \cdots\ \alpha_{M}\right]^{T} = D\alpha \end{aligned}$$
where $\alpha _{m}$ is a row vector based on sub-dictionary $D_{m}$ of the m-th class and $\left [\alpha _{\rm 1} \ \alpha _{\rm 2} \ \cdots \ \alpha _{\rm M}\right ]^{T}$ is a row vector too.

There are only non-zero values in $\alpha _{m}$ of $\left [\alpha _{\rm 1} \ \alpha _{\rm 2} \ \cdots \ \alpha _{\rm M}\right ]^{T}$ and the others are zero, so $\left [\alpha _{\rm 1} \ \alpha _{\rm 2} \ \cdots \ \alpha _{\rm M}\right ]^{T}$ is sparse. For approximating sample $x$ with dictionary $D$ and coefficient $\alpha$, an optimization problem is constructed, which is regularized by the 0-norm of coefficient $\alpha$.

$$\min_{\alpha}{\left\|x-D\alpha\right\|_{2}^{2}+\lambda\left\|\alpha\right\|_0}$$

The optimization problem is solved to obtain the sparse vector $\alpha$. Then sample $x$ is classified.

$$Class(x) = \min_{m = 1,\ldots,M}{\left\|x-D\alpha_{m}\right\|_{2}}$$
where $Class(x)$ is the classifying result.

2.2 Nonlinear sparse representation classification

With the in-depth study of SRC, some scholars hold that the SRC of high-dimensional data can not be simply considered as a linear one. To solve this, a kernel sparse representation classifier (KSRC) is proposed by mapping the samples into kernel space via kernel tricks to capture their nonlinear structure [24], which makes the data easier to group. Gan L et al. present a multiple-feature kernel sparse representation classifier (MFKSRC) to improve the classification effect of the model by combining the kernel method with SRC [25]. As the research progresses, they found that only using a single kernel function cannot be reflected in classification. For avoiding the above mentioned issues, Dan Li et al. propose a novel adaptive kernel sparse representation method based on multiple feature learning (AKSR-MFL) with multi-kernel functions [26]. Sample $x$ and dictionary $D$ are converted into kernel space to construct their optimization model as fellows.

$$\min_{\alpha}{\left\|\Phi(x)-\Phi(D)\alpha\right\|_{2}^{2}+\lambda\left\|\alpha\right\|_0}$$
where $\Phi (.)$ is a nonlinear mapping that transfers the original space to a high-dimensional kernel space. Then sample $x$ is classified.
$$Class(x) = \min_{m = 1,\ldots,M}{\left\|\Phi(x)-\Phi(D_{m})\alpha_{m}\right\|_{2}}$$

2.3 New style of SRC research

Its time complex is so high that $\Phi (.)$ converts the high-dimensional $x$ and $D$ to a high-dimensional kernel space in Eq. (6). For high performance, some scholars focus on the correlation between similar samples, especially, their approximate linear relations in low-dimensional subspaces. Chen et al. are interested in the similarity of neighbor samples and unveil two studying paths [27]. One is that they pay close attention to the contextual information that comes from a linear combination of four neighbors of a test when obtaining sparse coefficient vectors. Another is that they exploit the correlation between neighboring tests to construct an adopting joint sparsity model, where tests in a small neighborhood are assumed to be represented by a few common training samples. The second path is named joint sparsity representation.

Joint sparsity representation is a prominent direction for SRC research now. Zhang et al. assume that all neighbors of the central test share the same atoms and draw joint sparse representation classification(JSRC) [28]. And Peng et al. improve its performance in dictionary construction and sparse representation to set up local adaptive joint sparse representation (LAJSR) [29]. Recently, Zhang et al. expand the neighbor sample to a 7 $\times$ 7 window and assign different weights to its neighbor sample to realize weight joint sparse representation classification [30]. It constructs the joint sample matrix $X = \left [x_1 \ x_2 \cdots x_{49} \right ]$ of the 7 $\times$ 7 window, whose first column is the test located in the central test of window and the rest of the columns are its neighbors that are placed randomly in $X$. It is named as nonlocal weight joint sparse representation classification (NLW-JSRC) because the window of neighbor samples is extended to 7 $\times$ 7. So, NLW-JSRC is described as fellows.

$$\min_{A}{\left\|XW-DA\right\|_{2}^{2}+\lambda\left\|A\right\|_0}$$
where $A = \left [\alpha _1 \ \alpha _2 \cdots \alpha _{49} \right ]$ is a coefficient matrix. $W$ is a weight matrix.

Solving the optimization objective of NLW-JSRC, the test $x$ can be labeled by the minimal residual.

$$Class(x) = \min_{m = 1,\ldots,M}{\left\|XW-D_{m}A_{m}\right\|_{2}}$$

There are non-uniform regions and the adjacent samples in a window may be usually imaged by different features in UAV photographs, which contradicts the correlation assumptions of JSRC. In addition, the coefficients are different, and samples in a window are fitted by the same atoms. JSRC needs to be improved.

And it is undeniable that RGB images can form a high-dimensional space after extracting features, and sparsity also exists within any high-dimensional space. However, there is still a vacancy in SRC for cropland delineation with the high-dimensional data extracted from RGB photographs. Thus, we focus on the SRC method of cropland delineation from RGB UAV photographs.

3. Proposed methodology

3.1 Framework of the methodology

Because there is no complete spectral information in RGB UAV photographs, extracting feature information to construct an original high-dimensional dictionary is the primary task of sparse representation classification. Then the original dictionary is trained by online dictionary learning for sparse coding (ODL) [31] and the atoms are obtained to compose classification space and a multi-feature association dictionary (MFD). After that, the neighboring samples of the same feature structures are adaptively found based on a test via a correlation operator to make up a window. Finally, a constraint term is built through the similarity relationship of neighbors in the window to reconstruct the optimization objective of SRC. Leveraging the aforementioned ideas, the framework of our algorithm, multi-feature sparse representation based on adaptive graph constraint (AMFSR), is shown in Fig. 1.

 figure: Fig. 1.

Fig. 1. Framework of AMFSR.

Download Full Size | PDF

3.2 Multi-feature association dictionary

The relatively low number of spectral bands means that algorithms cannot obtain sufficient information from RGB aerial images, so the mission of its delineation remains challenging. To get more characteristic information from the three channels, we extract the color characteristics of each channel in different color spaces and settle for 12 characteristics to reflect the original spectrum from different perspectives. The features are combined to replace the original RGB space, which forms a new multi-feature space for multi-feature cropland identification.

The features are highly correlated with each other, it is not reasonable to establish a sparse representation classifier in each feature space, independently. Moreover, the dimension of feature vectors extracted from each feature space is different. Forcing the feature vectors to the same dimension will lose information and lead to misclassification [32]. So the multi-feature association dictionary (MFD) is propose to combine the sparse representation models of diverse features into an SRC mode. We describe the i-th feature sub-dictionary as $D_{i} \left (i = 1, \ldots,N\right )$. $N$ is the number of features and it is set to 12 in this paper. Relevantly, sample $x$ is defined as a vector $[x_1 \cdots x_i \cdots x_N ]^T$ that consists of feature data $x_i$. The i-th feature of the m-th class is denoted as $D_{i,m}(m=1, \ldots,M)$. $M$ is the number of sample categories and $D_i=\left [D_{i,1} \cdots D_{i,m} \cdots D_{i,M} \right ]$.

$$MFD=diag\left(D_{1},D_{2},\ldots,D_{i},\ldots,D_{N}\right)$$

Thus, multi-feature association sparse representation (MF-SR) for a single test can be described as fellows.

$$\min_{\alpha_{1},\alpha_{2},\ldots,\alpha_{N}}{\left\| \left [ \begin{matrix} x_{1} \\ \vdots \\ x_{i} \\ \vdots \\ x_{N} \end{matrix} \right ] - \left [ \begin{matrix} D_{1} & & & & \\ & \ddots & & & \\ & & D_{i} & & \\ & & & \ddots & \\ & & & & D_{N} \end{matrix} \right ] \left [ \begin{matrix} \alpha_{1} \\ \vdots \\ \alpha_{i} \\ \vdots \\ \alpha_{N} \end{matrix} \right ] \right\|_{2} + \lambda\sum_{i=1}^N\left\|\alpha_{i}\right\|_0 }$$
where $\alpha _i$ in $\alpha =\left [\alpha _{1} \cdots \alpha _{i} \cdots \alpha _{N} \right ]^T$ is the coefficient vector of i-th feature and $\alpha _{i,m}$ in $\alpha _i=\left [\alpha _{i,1} \cdots \alpha _{i,m} \cdots \alpha _{i,M} \right ]$ is the coefficient of $D_{i,m}$.

3.3 Shape-adaptive window

Descriptions of the hills and mountains are different in UAV photographs. Cropland is scattered and its shape is irregular. Consequently, the algorithms using a rectangular window, such as NLW-JSRC, to obtain spatial structure are invalid. The irregular shape windows of similar spatial structures need to be adaptively constructed, which satisfies that the samples in a shape window are similar or the same class.

Drawing support from the idea of breadth-first search, the algorithm for constructing the shape window is designed. The algorithm is a searching algorithm. A test block (such as a cropland block) is regarded as a primary center sample $I_{center}$ and the starting point of the algorithm. A neighbor $I_j$ of $I_{center}$ is searched, and it will be added to window $T$ if the neighbor $I_j$ satisfies the conditions that it has not been searched and has a strong correlation with the center sample $I_{center}$. This process is iteratively executed as Fig. 2.

 figure: Fig. 2.

Fig. 2. Process to build an adaptive window of a test in Algorithm 1.

Download Full Size | PDF

For controlling size of a window and ensuring that the samples in a window are similar, a similarity measure $\omega$ is defined as fellows.

$$\omega_{i,j}=\frac{exp(-\frac{\left\|J(x^i)-J(x^j)\right\|_{2}}{\sigma^2})}{\sqrt{(locx_i-locx_j)^2+(locy_i-locy_j)^2}}$$
where $\sigma$ is the standard deviation of the sample in the dictionary. In the numerator, $x^i$ is characteristic information of the i-th sample. The denominator is the Euclidean distance between a center sample and its neighbor (a sample).

Equation (11) manifests distance between a center sample and its neighbor is smaller, and $\omega$ is bigger. Meanwhile, the similarity of a center sample and its neighbor is higher, and $\omega$ is bigger too. Defining only the $\omega$ measure cannot control the growth of the window, and a threshold $\xi$ needs to be set Newly added neighbors will be changed for the center sample and the above operations are performed iteratively until $\omega < \xi$ . A shape window will be built after all iterations based on a test block are done.

The process of building a shape window is summarized as Algorithm 1.

Tables Icon

Algorithm 1. Adaptive window construction algorithm based on breadth-first search

3.4 AMFSR

3.4.1 Improving MF-SR with adaptive graph constraint

After window T is built, samples in it are akin and share similar sparse patterns. However, the characteristic values of each sample are uneven, which means the representation coefficients of the same atom for each sample are different. To limit the difference of their coefficients, the weight that is calculated with Eq. (11) is introduced to construct an adaptive graph constraint and improve the optimization objective of MF-SR, which is named multi-feature sparse representation based on adaptive graph constraint (AMFSR). Thus, the optimization objective of AMFSR is established as follows.

$$\min_{\alpha^{1},\alpha^{2},\ldots,\alpha^{K}}{\sum_{i=1}^K\left\|x^i-MFD\alpha^{i}\right\|_{2}^2 + \lambda\sum_{l=1}^K\left\|\alpha^{l}\right\|_0+\sum_{i,j}\left(\alpha^{i}-\alpha^{j}\right)\omega_{i,j} }$$
where $\alpha ^i$ is coefficient vector of $x^i$ and $K$ is the number of samples in T.

A regularization term, $\sum _{i,j}\left \|\alpha ^{i}-\alpha ^{j}\right \|\omega _{i,j}$, is implanted into Eq. (12). When $\left \|\alpha ^{i}-\alpha ^{j}\right \|$ is bigger, a smaller $\omega _{i,j}$ reconciles that $\sum _{i,j}\left \|\alpha ^{i}-\alpha ^{j}\right \|\omega _{i,j}$ is smaller. When $\omega _{i,j}$ is smaller, the difference $\left \|\alpha ^{i}-\alpha ^{j}\right \|$ is passed. The reason is that $x^i$ and $x^j$ are 2 different samples and their coefficients should be unequal as well. When the optimization objective is minimized, this regularization term satisfies the requirement that the coefficient is various based on sharing the same atoms.

3.4.2 Solving the optimization problem

(1) Matrix form of Eq. (12)

Generally, $\left \|\alpha ^{l}\right \|_0$ will be converted to $\left \|\alpha ^{l}\right \|_1$ for solving each $\alpha ^l$ in Eq. (12) by the feature-sign search algorithm [33]. However, it is not suitable for large-scale matrix operations. For ease of computation, a Laplace matrix $L$ is defined for each test.

$$L=U-W,U=diag\left(u_1,u_2,\ldots,u_K\right),u_i=\sum_{j=1}^K\omega_{i,j}$$
where $W$ is a matrix of $\omega _{i,j}$.

$L$ is a measure based on feature differences and spatial distances. So, Eq. (12) is converted into matrix form.

$$\min_{A}{\left\|X-MFDA\right\|_{2}^2 + \lambda\left\|A\right\|_{low,0}+tr\left(ALA^T\right) }$$
where $X$ is the matrix of samples in window T and $A$ is its coefficient matrix.

Then, in each sub-dictionary, the residual difference between the reconstructed data of the sample and the original data is used to realize the classification.

$$Class(x)=\min_{m=1,\ldots,M}{\left\|X-D_{m}A_{m}\right\|_{2} }$$
where $x$ is the sample to be tested and $D_m=\left [D_{1,m} \ D_{2,m} \cdots D_{N,m}\right ]$.

(2) Solving $A$ of Eq. (14)

Because samples $X$ can be fitted by the dictionary $MFD$ and it is $X = MFDA$, the following process holds true.

$$X=MFDA \Rightarrow XA^T=MFDAA^T \Rightarrow MFD= X A^T (AA^T )^{{-}1}$$
where $A$ is a multidimensional coefficient matrix and $AA^T$ is a nonsingle value matrix obviously.

To define $\widetilde {A} =A^T (AA^T )^{-1}$, thus$MFD=X\widetilde {A}$ and the objective function is modified as fellows.

$$\min_{\widetilde{A}}{\left\|MFD-X\widetilde{A}\right\|_{2} + \lambda\left\|\widetilde{A}\right\|_{low,0}+tr(\widetilde{A}^{T}L\widetilde{A}) }$$

The objective function of Eq. (17) was derived and its derivation is set to zero. The optimal solution is

$$\widetilde{A}=(X^{T} X+L)^{{-}1} X^T MFD$$

According to the matrix operation is easy to draw the relationship between $\widetilde {A}$ and $A$, their relationship proof is

$$\widetilde{A}=A^{T} (AA^{T} )^{{-}1} \Rightarrow \widetilde{A}AA^{T}=A^{T} \Rightarrow \widetilde{A}AA^{T} A=A^{T}A \Rightarrow \widetilde{A}A=E$$
(3) Iterative process of solving A

Equation (19) cannot satisfy the constraint of $\lambda \left \|\widetilde {A}\right \|_{row,0}$, which can cause that analytical solution may not be sparse. The idea is based on [33], a greedy algorithm is re-designed to approximately solve the objective Eq. (17). Firstly, the correlation matrix $W$ and window T are obtained with Algorithm 1 for each test, and the times h of iteration is set to 1, and the optimal dictionary $D_0$ is $\Phi$ and $R_{h-1}$ is initialized to the samples in window T. Each atom $d_k$ in the dictionary $MFD$ is sequentially injected into the following Eq. (20).

$$\mu=\max_{k=1,\ldots,K}{\left\| R_{h-1}^Td_{k} \right\|_{2}}$$
where $K$ is the total number of atoms.

At each iteration, a suitable atom $d_\mu$ is selected to add into the optimal dictionary $D_h$, which can be best matched to the current residual $R_{h-1}$. So $D_h$ is

$$D_{h}=\left[D_{h-1} \ d_{\mu}\right]$$
where $d_{\mu }$ is $\mu$-th atom of $MDF$.

When current $D_{h}$ is settled in one cycle iteration, $\widetilde {A}_{h}$ is solved with Eq. (18) and $A_{h}$ is obtained with Eq. (19). The residual $R_{h}$ is updated with the following formula.

$$R_{h}=X-D_{h}A_{h}$$

Then, the next iteration is executed until the times of iteration h satisfy the sparsity parameter H. The reconstruction coefficient $A$ is $A_H$ and Atoms in the optimal dictionary $D_h$ are greedily selected, which can approximate the original sample matrix of window T.

(4) AMFSR

Above process is summarized as Algorithm 2.

3.5 Time cost analysis

There are two main steps in AMFSR. The first is to build an adaptive window for each test. Its process is to find out all satisfying samples from the near neighbors of the test with breadth-first search and similarity threshold. So, its time complexity is O(2N-1) according to the time complexity of breadth-first search and N is the number of samples that enter into the adaptive window. Its meaning is that a sample search is performed 2N-1 times. The second is to reconstruct the spare coefficients. Its loop body is to optimize the dictionary and solve the spare coefficients. The sparsity parameter H determines the times that its circle is executed. Its time complexity based on the loop is O(H).

For an Image with $L\times W$ and size of test is $l\times l$, the number of tests will be $(L\times W)/(l\times l)$. Therefore, time cost of AMFSR is to set up $(L\times W)/(l\times l)$ adaptive windows and the cycle to reconstruct the spare coefficients is done $(L\times W)/(l\times l)$ times.

4. Experimental results and analysis

4.1 Experimental area and data collection

This research was in the Bishan and Hechuan districts located in Chongqing, China. The topography of the study area is mostly mountainous and hilly and the land cover type of them is agricultural. The cropland area of the Bishan district is about 281.8601 square kilometers. Hechuan is known as the ’Grain Silo of Bayu’, whose cropland area is about 976 square kilometers.

In 2022, the data sets are captured from rural areas in the Bishan and Hechuan districts by UAV. We collected RGB images using the Zenith P1 camera mounted on a DJI M300RTK. The shooting parameters are in Table 1.

An overview of the experimental area for this study is shown in Fig. 3. And 5 RGB images were randomly selected to verify the performance of the algorithm.

 figure: Fig. 3.

Fig. 3. Overview of the experimental area of this study. (a) The experimental areas are located in the Bishan and Hechuan districts of Chongqing, China. (b) UAV photograph of the experimental area in Hechuan. (c) Data captured in Hechuan. (d) UAV photograph of the experimental area in Bishan. (e) Data captured in Bishan.

Download Full Size | PDF

Data 1: Its size is $620\times 1240$ pixels and it is extracted from Fig. 3 (d).

Data 2: Its size is $620\times 1240$ pixels and it is extracted from Fig. 3 (d).

Data 3: Its size is $360\times 780$ pixels and it is extracted from Fig. 3 (b).

Data 4: Its size is $400\times 780$ pixels and it is extracted from Fig. 3 (b).

Data 5: Its size is $400\times 780$ pixels and it is extracted from Fig. 3 (b)

4.2 Experimental design and environment

4.2.1 Experimental design

To verify the effectiveness of the proposed algorithm, six experiments were designed.

Experiment 1: It is the experiment of selecting AMFSR parameters, which are sparsity $H$ and similarity threshold $\xi$. AMFSR identifies cropland with Data 1 to compare their recognition accuracy under the different parameters.

Experiment 2: It is the experiment of selecting sample sizes. AMFSR identify cropland with the Data 1, Data 2 and Data 3 to compare their recognition accuracy under the different sample sizes.

Experiment 3: It is a comparative experiment on single-feature dictionaries and a multi-feature association dictionary. AMFSR identifies cropland with Data 1 and Data 2 to compare their recognition accuracy under the different dictionaries.

Experiment 4: It is an accuracy comparative experiment with Data 1 between AMFSR and the comparative algorithms that involve SVM [6], LRC [9], KNN, SP, DeeplabV3+,and NLW-JSRC [30].

Experiment 5: It is the repeating experiment of Experiment 3 and Experiment 4 with Data 3, which tests the adaptability of AMFSR.

Experiment 6: It is a generalization ability experiment of AMFSR. AMFSR is trained by Data 3 to identify the cropland in Data 4 and Data 5, and their recognition accuracy is compared.

4.2.2 Experimental environment

This experiment is based on Intel Xeon silver 4114 CPU @ 2.20GHz (2 processor), 64GB memory, NVIDIA Titan V graphics card, windows 10 professional workstation 64 bit, Python 3.6.8, and PyCharm 2021.1.1(Professional Edition).

4.3 Experiment results and analysis

4.3.1 Analysis of parameters

Setting different sparsity $H$ and similarity threshold $\xi$ parameters, and the overall accuracy (OA) of Experiment 1 is shown in Fig. 4.

 figure: Fig. 4.

Fig. 4. OA of Experiment 1 with the varying $H$ and $\xi$ parameter. (a) OA based on the varying $H$ parameter. (b) OA based on the varying $\xi$ parameter.

Download Full Size | PDF

Figure 4 shows that OA increases along with the increase of $H$ when $\xi$ is fixed and $H$ is relatively small. When the sparsity parameter $H$ increases to a certain extent and the atomic number satisfies the requirement that the sample can be reconstructed by the atom, OA of the recognition task will arrive at the highest point. Continuing to increase $H$ beyond the optimal value, OA of the recognition task cannot continue to be improved and even will be decreased due to the introduction of atoms from different classes.

When $H$ is fixed and $\xi$ increases, more similar samples are selected to constrain the test, which improves the accuracy of cropland identification. Then $\xi$ continues to go up, and the adaptive window shrinks to exclude the more relevant samples, which eventuates misclassification and OA of AMFSR will be decreased. Figure 4 indicates that $H=2400$ and $\xi =0.95$ are the better parameters. So, they are chosen as parameters for subsequent experiments.

4.3.2 Results and Analysis of Experiment 2

Experiment 2 is designed to explore the effect that is caused by the sample block size on iteration time and recognition accuracy. Sample sizes of $5\times 5$ px, $10\times 10$ px, $15\times 15$ px, $20\times 20$ px, $25\times 25$ px, $30\times 30$ px, $50\times 50$ px, and $100\times 100$ px will be taken, respectively. The image is segmented into samples of different sizes, then 10% of the samples is the training samples and the rest are test. AMFSR will be executed with the different size samples and their OA and iterations are shown in Fig. 5.

 figure: Fig. 5.

Fig. 5. The Impact of different size samples. (a) OA of different size samples. (b) The iteration time of different sample sizes.

Download Full Size | PDF

Figure 5 presents that OA tends to increase as the sample size increases, when the sample size of AMFSR is smaller than 20$\times$20px. At this condition, the OA is mainly affected by the quantity of features included in a training sample. When the sample size is larger than 20$\times$20px, OA reduces along the sample size increase. The iterating times of AMFSR will be decreased while the sample size increases. The highest recognition rate is achieved for identifying cropland with a sample size of 20$\times$20px, whilst the iteration of AMFSR is done 1922 and 1035 times in our experiment respectively.

More iteration times of AMFSR are required for excessively small size of the samples. The larger the size of the sample is set and the fewer samples can be segmented from the image, thereby the iteration time is reduced. However, as the sample size is increased, the different features will be divided into a block, which leads to a decrease in OA of the AMFSR. Therefore, the block of 20$\times$20 px is the better sample and it is chosen as the size of subsequent experiment samples.

4.3.3 Results and analysis of experiment 3

The histograms of RGB, HSV, and Lab as the dictionary of RGB (RGB-D), the dictionary of HSV (HSV-D), and the dictionary of Lab (Lab-D) are extracted respectively. The multi-feature association dictionary (MFD) is obtained by associating the features of the three color spaces. AMFSR recognizes cropland with these different dictionaries and the results are shown in Fig. 6 and Fig. 7.

 figure: Fig. 6.

Fig. 6. The image results of Experiment 3 with different dictionaries on Data 1. (a) Original image of Data 1. (b) Label image of Data 1. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.

Download Full Size | PDF

 figure: Fig. 7.

Fig. 7. The image results of Experiment 3 with different dictionaries on Data 2. (a) Original image of Data 2. (b) Label image of Data 2. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.

Download Full Size | PDF

In Table 2, Acc_Cro = (Cro+)/( Cro_all+ Oth2Cro). Cro+ is number of samples that cropland in data sets is correctly identified with AMFSR. Cro_all indicates number of samples that cropland in data sets Oth2Cro means number of samples that non-cropland in data sets is identified as cropland with AMFSR. Acc_Oth is defined as (Oth+)/( Oth_all+ Cro2Oth). Oth+ describes number of samples that non-cropland in the data sets is correctly identified with AMFSR. Oth_all and Cro2Oth are similar to Cro_all and Oth2Cro respectively. And Acc_Cro and Acc_Oth are same in subsequent experiments.

Tables Icon

Table 1. The shooting parameters.

Tables Icon

Table 2. Recognition Accuracy of AMFSR with different dictionaries on Data 1.

Acc_Of_Cro of AMFSR that exploits RGB-D is only 3.8845%, while the recognition accuracy can reach 83.6000% with MFD in Table 2. For the cropland of Data 2, Acc_Of_Oth of AMFSR with Lab-D is only 9.5859%, while the recognition accuracy of the classifier can increase to 80.7531% with MFD. Kappa of AMFSR with MFD are all greater than 0.8 in the above experiments, and their accuracies of identifying cropland and others are almost identical. The aforementioned result proves that the joint dictionary of multiple features improves recognition accuracy, which is shown in Table 2 and Table 3.

Tables Icon

Table 3. Recognition Accuracy of AMFSR with different dictionaries on Data 2.

4.3.4 Results and analysis of experiment 4

The image results of Experiment 4 are shown in Fig. 8.

 figure: Fig. 8.

Fig. 8. The image results of Experiment 5 with AMFSR and its Comparison algorithms on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) AMFSR. (d) NLW-JSRC. (e) DeeplabV3+. (f) SP. (g) SVM. (h)LRC. (i)KNN.

Download Full Size | PDF

The parameters of AMFSR are $\xi =0.95$ and H=2400. The parameters of NLW-JSRC algorithm are $H=2400$ and the window size is $7\times 7$. The sigmoid function is chosen as the kernel function in SVM. The L2 regularization with coefficient $\lambda =0.01$ is used in LRC. The parameter of KNN is 3. The backbone of deeplabV3+ is deeplabv3_resnet50. Their parameters are same in subsequent experiments as well.Experiment 4 is done ten times repeatedly and averages of their recognition accuracies are compared in Fig. 9.

 figure: Fig. 9.

Fig. 9. Recognition Accuracy of AMFSR and comparison algorithms on Data 1 with different atom numbers. (a) Acc_Cro of AMFSR and comparison algorithms. (b) Acc_Oth of AMFSR and comparison algorithms. (c) OA of AMFSR and comparison algorithms. (d) Kappa of AMFSR and comparison algorithms.

Download Full Size | PDF

As shown in Fig. 9, the number of atoms can affect OA. The recognition accuracy of the seven algorithms is gradually improved along with increasing the number of atoms. AMFSR, NLW-JSRC and DeeplabV3+ have significant recognition results, and DeeplabV3+ is superior to the other compared algorithms. The highest accuracy of DeeplabV3+ is around 80% and the overall accuracy for AMFSR is 86.0810%. Even compared with NLW-JSRC and DeeplabV3+, the accuracy of the proposed method is 6.2076% and 6.0810% higher.

In the single-image recognition task, OA of DeeplabV3+ is slightly lower than OA of AMFSR. The performance of DeeplabV3+ for large crop-area is comparable to the performance of AMFSR, and it is worse than AMFSR for dispersed and fragmented areas. Due to NLW-JSRC considering the connections of the adjacent samples, its performance is superior to SVM. In contrast to NLW-JSRC, AMFSR not only performs similarity-based selection of near-neighbor samples but also allows the coefficients of similar samples to be varied. So OA of AMFSR is greater than NLW-JSRC.

4.3.5 Results and analysis of experiment 5

To test the adaptability of AMFSR, Data 3 is utilized to repeat Experiment 3 and Experiment 4. The parameters are same as Experiment 3 and Experiment 4. First, AMFSR with different dictionaries is applied to recognize cropland and the results are shown in Fig. 10 and Table 4. Acc_Oth of AMFSR with RGB-D is only 1.3890% on Data 3, while the recognition accuracy of the MFD-based algorithm can reach 85.5478%. The accuracy of Data 3 is slightly higher than that of Data 1 and Data 2.

 figure: Fig. 10.

Fig. 10. The image results of Experiment 5 with different dictionaries on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.

Download Full Size | PDF

Tables Icon

Table 4. Recognition Accuracy of AMFSR with different dictionaries on Data 3.

Then, the six algorithms in Experiment 5 are compared with AMFSR and Experiment 5 is done ten times with Data 3. Their image results and averages of accuracies are presented in Fig. 11 and Fig. 12 repeatedly.

 figure: Fig. 11.

Fig. 11. The image results of Experiment 5 with AMFSR and its Comparison algorithms on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) AMFSR. (d) NLW-JSRC. (e) DeeplabV3+. (f) SP. (g) SVM. (h)LRC. (i)KNN.

Download Full Size | PDF

 figure: Fig. 12.

Fig. 12. Recognition Accuracy of AMFSR and comparison algorithms on Data 1 with different atom numbers. (a) Acc_Cro of AMFSR and comparison algorithms. (b) Acc_Oth of AMFSR and comparison algorithms. (c) OA of AMFSR and comparison algorithms. (d) Kappa of AMFSR and comparison algorithms.

Download Full Size | PDF

As the number of atoms increases, the recognition effect of AMFSR improves. When the atom number reaches the number of samples that can be reconstructed, the recognition accuracy of AMFSR is better than its comparison algorithm. OA of AMFSR is 91.4432%, which is 18.5241% higher than NLW-JSRC. The experiment proves that the AMFSR is more effective in recognizing regions of concentrating cropland and its adaptability is strong. DeeplabV3+ is more accurate than the other comparison algorithms. However, based on our single-image cropland recognition task, its recognition accuracy is also lower than AMFSR.

4.3.6 Results and analysis of experiment 6

The left subgraph (Data 4) and right subgraph (Data 5) of Data 3 are chosen as the data for the generalization experiment. AMFSR_1 based on AMFSR is trained by the training samples from Data 3 and AMFSR_2 based on AMFSR is built by the training samples from Data 3 and its subgraphs. The cropland in Data 4 and Data 5 is recognized with AMFSR_1 and AMFSR_2 respectively, and their recognition accuracy compared to evaluate the generalization ability of AMFSR.

The image results with AMFSR_1 and AMFSR_2 are shown in Fig. 13 (a) and Fig. 13 (b). Their recognition accuracy is listed in Table 5.

 figure: Fig. 13.

Fig. 13. The image results of Experiment 6 with different trained AMFSR on the Data 4 and Data 5. (a) Recognition results with AMFSR_1. (b) Recognition results with AMFSR_2.

Download Full Size | PDF

Tables Icon

Table 5. Recognition Accuracy of different trained AMFSR on Data 4 and Data 5.

Figure 13 and Table 5 show that OA of AMFSR_2 is 89.3618% on Data 4 and it is 4.7359% higher than that of AMFSR_1. OA of AMFSR_2 is 90.7418% on Data 5 and it is 1.9696% higher than that of AMFSR_1. Several conclusions can be made from Fig. 13 and Table 5. The cropland features of Data 5 are very similar to those of Data 3, so the generalization ability of Data 3 trained mode is superior in Data 4. While the cropland features of Data 4 are different from Data 3, thereby generalization ability of the model is poor. AMFSR_2 incorporates characteristics of cropland in Data 4, and its OA of Data 4 is significantly improved. Hence, the generalization ability of our proposed AMFSR model depends on the features of training samples in the dictionary. In other words, if the AMFSR model dictionary contains the characteristics of the current cropland, the model can be extended it to identify arable land. Otherwise, the recognition accuracy will be reduced.

5. Conclusion

A new sparse representation classification is proposed for the cropland delineation, which is a multi-feature sparse representation based on adaptive graph constraint (AMFSR). A multi-feature association dictionary is constructed with feature vectors that were extracted from aerial images. Then the adaptive window is established for each test sample, and graph regularization is built and a similarity weight measure is defined. The objective function based on MF-SR is reconstructed to obtain the optimization model of AMFSR.

  • 1. Constructing a multi-feature association dictionary (MFD) with the diagonal matrix effectively avoids the loss of feature data when they are abstracted as vectors. Based on MFD, the multi-feature association sparse representation (MF-SR) is proposed.
  • 2. An adaptive window is created to accommodate similar plot shapes. Its algorithm can adaptively explore the irregular spatial structure of cropland and others in aerial images.
  • 3. A weight based on features and Euclidean distance is defined to weight differences between the reconstructed coefficients of samples in a shape window, which can ensure that the representation coefficients of similar samples with the same atom are different. A graph regularization term is created with this weight and differences of the reconstructed coefficients to improve the optimization objective of MF-SR, which raises AMFSR. And it iterative solution algorithm is derived from the optimization model.
  • 4. The experimental results of the five data sets show that OA of AMFSR can reach 92.3546%, which is 14.0887% higher than NLW-JSRC. This proves that the multi-feature joint dictionary and adaptive graph regularity of AMFSR are effective.

Because the recognition sample granularity of AMFSR is coarse, a fine-grained algorithm for cropland recognition of RGB aerial images can be further researched and its recognition accuracy is yet to be improved in the future.

Funding

Natural Science Foundation of Chongqing (CSTB2022NSCQ-LZX0081); Chongqing Municipal Science and Technology Bureau (CSTB2022TFII-OFX0043); Chongqing Municipal Education Commission (No. CXQT20015).

Acknowledgments

This work supported by the Key Project of Natural Science Foundation (CSTB2022NSCQ-LZX0081) of Chongqing, Technology Foresight and Institutional Innovation Project (CSTB2022TFII-OFX0043) of Chongqing, and Chongqing University Innovation Research Group Funding (No. CXQT20015) of Chongqing Municipal Education Commission, China.

Disclosures

The authors declare no conflicts of interest.

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

References

1. T. Feng, R. Xiong, and P. Huan, “Productive use of natural resources in agriculture: The main policy lessons,” Resour. Policy 85, 103793 (2023). [CrossRef]  

2. Y. Chen, R. Xin, and H. Jiang, “Refined feature fusion for in-field high-density and multi-scale rice panicle counting in uav images,” Comput. Electron. Agric. 211, 108032 (2023). [CrossRef]  

3. A. Guo, W. Yue, and J. Yang, “Cropland abandonment in china: Patterns, drivers, and implications for food security,” J. Cleaner Prod. 418, 138154 (2023). [CrossRef]  

4. P. Singh, P. C. Pandey, G. P. Petropoulos, et al., “8 - hyperspectral remote sensing in precision agriculture: present status, challenges, future trends,” in Hyperspectral Remote Sensing, P. C. Pandey, P. K. Srivastava, H. Balzter, et al., eds. (Elsevier, 2020), Earth Observation, pp. 121–146.

5. H. Wang, Z. Ye, and Y. Wang, “Improving the crop classification performance by unlabeled remote sensing data,” Expert Syst. with Appl. 236, 121283 (2024). [CrossRef]  

6. L. Cai, H. Wang, and Y. Liu, “Is potential cultivated land expanding or shrinking in the dryland of china? spatiotemporal evaluation based on remote sensing and svm,” Land Use Policy 112, 105871 (2022). [CrossRef]  

7. M. Belgiu and L. Dragut, “Random forest in remote sensing: A review of applications and future directions,” ISPRS J. Photogramm. Remote. Sens. 114, 24–31 (2016). [CrossRef]  

8. F. Wang, X. Zhu, and X. Cheng, “Mmkdgat: Multi-modal knowledge graph-aware deep graph attention network for remote sensing image recommendation,” Expert Syst. with Appl. 235, 121278 (2024). [CrossRef]  

9. J. Zhou, S. Zeng, and B. Zhang, “Linear representation-based methods for image classification: a survey,” IEEE Access 8, 216645–216670 (2020). [CrossRef]  

10. A. Gkillas, D. Kosmopoulos, and K. Berberidis, “Cost-efficient coupled learning methods for recovering near-infrared information from rgb signals: Application in precision agriculture,” Comput. Electron. Agric. 209, 107833 (2023). [CrossRef]  

11. N. Akhtar and A. Mian, “Hyperspectral recovery from rgb images using gaussian processes,” IEEE Trans. Pattern Anal. Mach. Intell. 42(1), 100–113 (2020). [CrossRef]  

12. C. Zhang, J. Valente, and W. Wang, “Data on three-year flowering intensity monitoring in an apple orchard: A collection of rgb images acquired from unmanned aerial vehicles,” Data Brief 49, 109356 (2023). [CrossRef]  

13. T. Wang, B. Chen, and Z. Zhang, “Applications of machine vision in agricultural robot navigation: A review,” Comput. Electron. Agric. 198, 107085 (2022). [CrossRef]  

14. R. Wan Nurazwin Syazwani, H. Muhammad Asraf, M. Megat Syahirul Amin, et al., “Automated image identification, detection and fruit counting of top-view pineapple crown using machine learning,” Alex. Eng. J. 61(2), 1265–1276 (2022). [CrossRef]  

15. O. Ilniyaz, Q. Du, and H. Shen, “Leaf area index estimation of pergola-trained vineyards in arid regions using classical and deep learning methods based on uav-based rgb images,” Comput. Electron. Agric. 207, 107723 (2023). [CrossRef]  

16. B. Himadri, Z. Keqi, and R. M. S G. Daniel, “Delineation of tree patches in a mangrove-marsh transition zone by watershed segmentation of aerial photographs,” Remote Sens. 12(13), 2086 (2020). [CrossRef]  

17. H. Umut, S. Mamat, and C. Shuisen, “Estimating the leaf area index of winter wheat based on unmanned aerial vehicle rgb-image parameters,” Sustainability 11(23), 6829 (2019). [CrossRef]  

18. Y. Shao, J. Ren, and J. Campbell, “9.04 - multitemporal remote sensing data analysis for agricultural application,” in Comprehensive Remote Sensing, S. Liang, ed. (Elsevier, Oxford, 2018), pp. 29–38.

19. J. Cheng, H. Liu, and T. Liu, “Remote sensing image fusion via wavelet transform and sparse representation,” ISPRS J. Photogramm. Remote. Sens. 104, 158–173 (2015). [CrossRef]  

20. K. hua Hui, C. li Li, and L. Zhang, “Sparse neighbor representation for classification,” Pattern Recognit. Lett. 33(5), 661–669 (2012). [CrossRef]  

21. L. Yang, J. Chen, and R. Zhang, “Precise crop classification of uav hyperspectral imagery using kernel tensor slice sparse coding based classifier,” Neurocomputing 551, 126487 (2023). [CrossRef]  

22. K. Yan, Y. Xu, and X. Fang, “Protein fold recognition based on sparse representation based classification,” Artif. Intell. Med. 79, 1–8 (2017). [CrossRef]  

23. L. Xun, J. Zhang, and D. Cao, “Mapping cotton cultivated area combining remote sensing with a fused representation-based classification algorithm,” Comput. Electron. Agric. 181, 105940 (2021). [CrossRef]  

24. C. Zhao, W. Li, X. Li, et al., “Sparse representation based on stacked kernel for target detection in hyperspectral imagery,” Optik 126(24), 5633–5640 (2015). [CrossRef]  

25. L. Gan, J. Xia, P. Du, et al., “Multiple feature kernel sparse representation classifier for hyperspectral imagery,” IEEE Trans. Geosci. Remote Sensing 56(9), 5343–5356 (2018). [CrossRef]  

26. D. Li, Q. Wang, and F. Kong, “Adaptive kernel sparse representation based on multiple feature learning for hyperspectral image classification,” Neurocomputing 400, 97–112 (2020). [CrossRef]  

27. Y. Chen, N. M. Nasrabadi, and T. D. Tran, “Hyperspectral image classification using dictionary-based sparse representation,” IEEE Trans. Geosci. Remote Sensing 49(10), 3973–3985 (2011). [CrossRef]  

28. E. Zhang, X. Zhang, and L. Jiao, “Spectral-spatial hyperspectral image ensemble classification via joint sparse representation,” Pattern Recognit. 59, 42–54 (2016). [CrossRef]  

29. J. Peng, X. Jiang, N. Chen, et al., “Local adaptive joint sparse representation for hyperspectral image classification,” Neurocomputing 334, 239–248 (2019). [CrossRef]  

30. M. Imani and H. Ghassemian, “An overview on spectral and spatial information fusion for hyperspectral image classification: Current trends and challenges,” Inf. Fusion 59, 59–83 (2020). [CrossRef]  

31. S. Deng, B. Jing, and S. Sheng, “Impulse feature extraction method for machinery fault detection using fusion sparse coding and online dictionary learning,” Chin. J. Aeronaut. 28(2), 488–498 (2015). [CrossRef]  

32. L. Fang, C. Wang, S. Li, et al., “Hyperspectral image classification via multiple-feature-based adaptive sparse representation,” IEEE Trans. Instrum. Meas. 66(7), 1646–1657 (2017). [CrossRef]  

33. Y.-C. Song, F.-Y. Wu, Y.-Y. Ni, et al., “A fast threshold omp based on self-learning dictionary for propeller signal reconstruction,” Ocean Eng. 287, 115792 (2023). [CrossRef]  

Data availability

Data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (13)

Fig. 1.
Fig. 1. Framework of AMFSR.
Fig. 2.
Fig. 2. Process to build an adaptive window of a test in Algorithm 1.
Fig. 3.
Fig. 3. Overview of the experimental area of this study. (a) The experimental areas are located in the Bishan and Hechuan districts of Chongqing, China. (b) UAV photograph of the experimental area in Hechuan. (c) Data captured in Hechuan. (d) UAV photograph of the experimental area in Bishan. (e) Data captured in Bishan.
Fig. 4.
Fig. 4. OA of Experiment 1 with the varying $H$ and $\xi$ parameter. (a) OA based on the varying $H$ parameter. (b) OA based on the varying $\xi$ parameter.
Fig. 5.
Fig. 5. The Impact of different size samples. (a) OA of different size samples. (b) The iteration time of different sample sizes.
Fig. 6.
Fig. 6. The image results of Experiment 3 with different dictionaries on Data 1. (a) Original image of Data 1. (b) Label image of Data 1. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.
Fig. 7.
Fig. 7. The image results of Experiment 3 with different dictionaries on Data 2. (a) Original image of Data 2. (b) Label image of Data 2. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.
Fig. 8.
Fig. 8. The image results of Experiment 5 with AMFSR and its Comparison algorithms on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) AMFSR. (d) NLW-JSRC. (e) DeeplabV3+. (f) SP. (g) SVM. (h)LRC. (i)KNN.
Fig. 9.
Fig. 9. Recognition Accuracy of AMFSR and comparison algorithms on Data 1 with different atom numbers. (a) Acc_Cro of AMFSR and comparison algorithms. (b) Acc_Oth of AMFSR and comparison algorithms. (c) OA of AMFSR and comparison algorithms. (d) Kappa of AMFSR and comparison algorithms.
Fig. 10.
Fig. 10. The image results of Experiment 5 with different dictionaries on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) MFD. (d) RGB-D. (e) Lab-D. (f) HSV-D.
Fig. 11.
Fig. 11. The image results of Experiment 5 with AMFSR and its Comparison algorithms on Data 3. (a) Original image of Data 3. (b) Label image of Data 3. (c) AMFSR. (d) NLW-JSRC. (e) DeeplabV3+. (f) SP. (g) SVM. (h)LRC. (i)KNN.
Fig. 12.
Fig. 12. Recognition Accuracy of AMFSR and comparison algorithms on Data 1 with different atom numbers. (a) Acc_Cro of AMFSR and comparison algorithms. (b) Acc_Oth of AMFSR and comparison algorithms. (c) OA of AMFSR and comparison algorithms. (d) Kappa of AMFSR and comparison algorithms.
Fig. 13.
Fig. 13. The image results of Experiment 6 with different trained AMFSR on the Data 4 and Data 5. (a) Recognition results with AMFSR_1. (b) Recognition results with AMFSR_2.

Tables (7)

Tables Icon

Algorithm 1. Adaptive window construction algorithm based on breadth-first search

Tables Icon

Table 1. The shooting parameters.

Tables Icon

Table 2. Recognition Accuracy of AMFSR with different dictionaries on Data 1.

Tables Icon

Table 3. Recognition Accuracy of AMFSR with different dictionaries on Data 2.

Tables Icon

Table 4. Recognition Accuracy of AMFSR with different dictionaries on Data 3.

Tables Icon

Table 5. Recognition Accuracy of different trained AMFSR on Data 4 and Data 5.

Equations (22)

Equations on this page are rendered with MathJax. Learn more.

x d m 1 α m 1 + d m 1 α m 1 + + d m N m α m N m = [ d m 1   d m 2   d m N m ] [ α m 1   α m 2     α m N m ] T
x D 1 α 1 + D 2 α 2 + + D M α M = [ D 1   D 2     D M ] [ α 1   α 2       α M ] T = D α
min α x D α 2 2 + λ α 0
C l a s s ( x ) = min m = 1 , , M x D α m 2
min α Φ ( x ) Φ ( D ) α 2 2 + λ α 0
C l a s s ( x ) = min m = 1 , , M Φ ( x ) Φ ( D m ) α m 2
min A X W D A 2 2 + λ A 0
C l a s s ( x ) = min m = 1 , , M X W D m A m 2
M F D = d i a g ( D 1 , D 2 , , D i , , D N )
min α 1 , α 2 , , α N [ x 1 x i x N ] [ D 1 D i D N ] [ α 1 α i α N ] 2 + λ i = 1 N α i 0
ω i , j = e x p ( J ( x i ) J ( x j ) 2 σ 2 ) ( l o c x i l o c x j ) 2 + ( l o c y i l o c y j ) 2
min α 1 , α 2 , , α K i = 1 K x i M F D α i 2 2 + λ l = 1 K α l 0 + i , j ( α i α j ) ω i , j
L = U W , U = d i a g ( u 1 , u 2 , , u K ) , u i = j = 1 K ω i , j
min A X M F D A 2 2 + λ A l o w , 0 + t r ( A L A T )
C l a s s ( x ) = min m = 1 , , M X D m A m 2
X = M F D A X A T = M F D A A T M F D = X A T ( A A T ) 1
min A ~ M F D X A ~ 2 + λ A ~ l o w , 0 + t r ( A ~ T L A ~ )
A ~ = ( X T X + L ) 1 X T M F D
A ~ = A T ( A A T ) 1 A ~ A A T = A T A ~ A A T A = A T A A ~ A = E
μ = max k = 1 , , K R h 1 T d k 2
D h = [ D h 1   d μ ]
R h = X D h A h
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.