Toward the automatic detection of effective gerbil holes in desert grasslands through unmanned aerial vehicle imagery

Abstract Rapidly determining the density of effective gerbil holes is ecologically important but technically challenging. However, unmanned aerial vehicles (UAVs) offer new methods to identify effective gerbil holes and understand the spatial relationships between gerbils and grass coverage. In this study, we focused on Meriones unguiculatus, which live in desert grasslands of Ordos, and gathered UAV imagery to explore the performance of rule-based classification in object-based image analysis (OBIA) for identifying effective gerbil holes. To improve the classification accuracy and reduce the time cost of repeated ‘trial and error’, we propose a novel method to build rule sets. We adopted the estimation of scale parameter_2 (ESP2) method during the segmentation stage to enable fast and objective segmentation parameterization. For the classification stage, a correlation-based feature selection (CFS) algorithm was applied for feature selection and threshold prediction. Our analysis determined the following universally recommended rule sets for identifying effective gerbil holes using OBIA: scale parameter of 107, shape factor of 0.2, compactness value of 0.3, excess green value less than −5, brightness feature value in the 143–149 range and roundness feature less than 0.9. With these rule sets, OBIA exhibited higher classification accuracy than maximum likelihood (ML) classification, with an overall classification accuracy of 88.25%. We also found that the relationship between the area of effective gerbil holes (y) and the grass coverage (x) satisfied a quadratic function:y=−1.981x2+1.104x+0.416. This study provides practical guidance for grassland management and rodent infestation control.


Introduction
Inner Mongolia is home to the largest grassland in China, but severe and very severe desertification has affected a significant portion of the region, with 98% of desertification occurring in grasslands (Kang and Liu 2014;Guo et al. 2015).Despite efforts to control desertification over the past 50 years, the overall trend has not improved substantially (Wei and Fang 2009;Guo et al. 2014).Rodent populations have increased as a concomitant phenomenon of grassland desertification, leading to the destruction of bare soil structures, interruption of material and energy flow and exacerbation of the process of grassland desertification (L€ u et al. 2016;Tang et al. 2019).Rodents also gnaw on forage, reducing plant diversity (Chen et al. 2017).Mongolian gerbils (Meriones unguiculatus) are the hosts of the plague, and there has been a resurgence of plague outbreaks in the past seven years, with a direct correlation between outbreaks and the increased density of M. unguiculatus (Davis et al. 2008;Eads et al. 2016;Feng et al. 2020).Monitoring the density and distribution of gerbil populations, indicated by the number and area of effective gerbil holes, is crucial for grassland management and plague prevention and control (Liu et al. 2019).
To carry out rodent censuses, the traditional method is to count effective gerbil holes using ground-based field surveys.However, this method is laborious, lacks spatial information and carries a risk of acquiring plague infection.Additionally, existing data gaps for harsh environments further limit the accuracy of traditional methods (Tang et al. 2019).With the development of geospatial technologies, remote sensing has become an increasingly popular method to monitor rodent populations (He et al. 2013;Chen et al. 2017).Related studies have used the spectral characteristics of rodent hazard areas to estimate population density (Liu et al. 2013;Pang and Guo 2017).However, the low spatial resolution data used in these studies limits the efficiency of monitoring rodent population density.
In recent years, remote sensing has advanced significantly toward generating and applying high temporal and spatial resolution data (Bhardwaj et al. 2016;Van Cleemput et al. 2018).UAVs have been widely used in ecological research due to their ability to provide high-resolution data.For example, Dandois and Ellis (2013) used UAVs to monitor plant phenology in a 50 m Â 50m sample plot in Maryland, USA, while Sun et al. (2018) integrated UAV technology with a deep learning framework to extract pest areas in forest monitoring.The use of UAVs in animal behavior and population surveys has also attracted the attention of animal ecologists.Weissensteiner et al. (2015) used UAVs to assess the breeding behavior of birds and compared it to traditional survey methods, saving 85% of time.Rey et al. (2017) proposed a semi-automatic system with UAVs to detect animals in savannas and achieved 75% correct detection for a precision of 10%.In the field of environmental protection, Wu et al. (2020) used color features to recognize residual plastic film from UAV images, providing a decision-making basis for farmland environmental health assessment.Pi et al. (2021) constructed a low-altitude UAV platform and achieved high-precision classification of desertification degradation indicator plants, providing quantitative data for developing ecological restoration.
However, achieving high accuracy with high-resolution remote data has been challenging due to the heterogeneity in local pixel values and spectral similarities among different types of data.Processing remote sensing data with pixels as primary mapping units also poses a challenge (Dronova 2015;Ma et al. 2017a).
Object-based image analysis (OBIA) provides a solution for analyzing high-resolution images by offering meaningful semantic information and spatial characteristics between neighboring pixels (Dr agut ¸et al. 2010;Juan et al. 2017;Ma et al. 2021aMa et al. , 2021b)).Previous research on OBIA has focused on land-cover (Costa et al. 2014;Tehrany et al. 2014) and wetland (Dronova 2015) projects and the methods used are primarily based on decision trees (Laliberte et al. 2006;Mallinis et al. 2008), the K-nearest-neighbors algorithm (Tehrany et al. 2014;Ma et al. 2017b) and artificial neural networks (Marmanis et al. 2016;Zhou and Han 2021).However, the accuracy of these supervised classification methods is limited by data sources, training set size and segmentation parameters (Ma et al. 2017a;Chen et al. 2018).Additionally, selecting representative samples for each category is time-consuming and laborious and can result in significant human error.In contrast, unsupervised algorithms (Langner et al. 2014;Ma et al. 2014;Charoenjit et al. 2015) provide an automatic and flexible approach to remote sensing image classification by obtaining information from the spectrum, texture and geometry.Rule-based algorithms, as a representative of this approach, usually suggest an optimal parameter and require a threshold (Espindola et al. 2006;Johnson and Xie 2011;Dronova 2015) to perform unsupervised classification (Langner et al. 2014), which can be applied to a large amount of data from the same batch of flights, enabling real-time data processing.
Previous researches on rule-based algorithms for classification mostly dependent on subjective 'trial and error' methods, this leaded to uncertainty in classification accuracy (Johnson and Xie 2011;Belgiu and Dr agut ¸2014).ESP2 tool have been developed to promote automatic decision on segmentation specifications (Liu and Ren 2022).Related studies have shown that unsupervised segmentation based on ESP2 are, in contrast, less subjective and more time-efficient, making them suitable for use in operational remote imagery classification settings (Dr agut ¸et al. 2010(Dr agut ¸et al. , 2014)).Due to high-resolution remote sensing data provide more detailed information, there are more than 200 features per object after segmentation (Laliberte et al. 2012).The large number of features not only causes difficulties in the establishment of classification rules, but also easily leads to dimensional disasters.Feature selection enables to obtain data for comprehensive understanding of objects (Pal and Foody 2010), and, to identify representative features and make predictions about their thresholds (Ma et al. 2017a), has become an important process to improve classification accuracy and enhance classification efficiency.Ma et al. (2015) applied CFS method to implement dimensionality reduction of the object features prior to classification, and proposed that CFS is the optimal feature selection algorithm in OBIA.The related studies on feature selection have been conducted in supervised classification methods of OBIA.Thus, there are few research on rule-based algorithms classification in the OBIA field, and, there is no common consensus about the general effects of the combination of feature selection methods and unsupervised rule-based classification.
The main objective of this study is to develop a rule set for object-based unsupervised classification instead of relying on supervised classification methods in OBIA.To achieve this, we combined automatic segmentation scale estimation in multiresolution segmentation with the CFS method to predict segmentation parameters and feature threshold values.Our study focuses on the meaningful visual characteristics of effective gerbil holes in desert grasslands, building on previous OBIA research.To the best of our knowledge, this study is the first to use UAV imagery to identify effective gerbil holes, and the findings are essential for understanding high-resolution data in specific applications.

Study area
This study was conducted at Otog Banner (39 25'N, 107 58'E), Ordos Plateau.The Ordos Plateau is located southwest of the Inner Mongolia Autonomous Region, China, which is a typical representative area of the desert grassland of Inner Mongolia, with an approximate elevation of 1300 m (Figure 1).The mean annual temperature is 9.6 C, and the annual precipitation is approximately 237.4 mm, but the annual average evaporation is up to 2470 mm.The surface bare soil is mainly light chestnut and brown calcium bare soil, with a large sand content and poor loose nutrient characteristics (Liu et al. 2019).Otog Banner is one of the plague natural focuses of M. unguiculatus on the Inner Mongolian Plateau.

Acquisition of UAV imagery and sample selection
In this study, we collected aerial images of five sample plots with different densities of effective gerbil holes in April 2020 and April 2021.Each sample plot covered an area of 50 m Â 50 m.Table 1 provides information on the geographical distribution of the sample plots and the statistics of effective gerbil holes obtained through ground surveys.We used quadrotor M200 UAVs (DJI-Innovations, Shenzhen, China) as the data acquisition platform and selected clear and cloudless days from 12:00 to 14:00 to minimize the impact of shadows on the classification results.The altitude was set to 30 meters, and the forward and side overlaps were set to 70% and 80%, respectively.A total of 445 images were collected, with each sample plot having 89 images of size 5280 pixels Â 3956 pixels.We used Pix4D (https://www.pix4d.com/),a professional drone mapping software, to correct and mosaic the images and generate digital orthophoto maps (DOMs) with a resolution of 0.4 cm.The images were captured during the herbage re-greening period, when the leaves of forage were green, and the effective gerbil hole patches were clearly distinguishable from the grassland in terms of texture and color.The images also contained a significant amount of bare soil and a few black patches, which were formed by withered grass shadows and abandoned gerbil holes.However, these black patches did not affect the extraction of effective gerbil hole information.We classified the ground objects in the UAV imagery into three categories: grass patches, effective gerbil holes and soil patches.
To determine a universally applicable threshold for the extraction of effective gerbil holes, we randomly selected 60 object samples from each sample plot (20 from each category) for a total of 300 samples.We used 60% of the samples for training and the remaining 40% for testing.We created raster files of all the object samples in eCognition, and then, converted them into polygons using ArcGIS as the regions of interest (ROIs) for the maximum likelihood (ML) classification.

Segmentation scale estimation using ESP2
OBIA considers both spectral and morphological, contextual and proximity features in images; therefore, many previous studies have indicated that the OBIA method outperformed pixel-based classification in very-high-resolution images (Blaschke 2010;Zhu et al. 2020;Luo et al. 2021).The OBIA consists of two main function modules: image segmentation and target object extraction (Addink et al. 2007).Image segmentation is a process of dividing remote sensing images into discrete regions or objects that are homogeneous in terms of spatial or spectral characteristics.This process reduces the within-class spectral variation in the high spatial resolution imagery and the fragmentation rate of the object (Jing and Cheng 2012).The OBIA in this study is performed using eCognition Developer 64 software, which is the world's first object-based classification software (Dr agut ¸et al. 2010(Dr agut ¸et al. , 2014)).
The commonly used segmentation methods in eCognition Developer include the chessboard, quadtree, contrast, multiresolution, spectral difference, multithreshold and contrast filter segmentation methods.Among them, multiresolution segmentation, which uses the region-growing segmentation algorithm, is the most effective segmentation method for acquiring information at different scales (Zhang et al. 2010;Mengmeng and Noboru 2017).For multiresolution segmentation, the most critical procedure is to choose appropriate segmentation parameters (typically containing scale parameters, shape factors and compactness values) to increase the interclass spectral differences of images and improve the accuracy of classification and identification (Addink et al. 2007).However, it is challenging to define the optimal segmentation parameters for image objects of different sizes and shapes in different scenes of remote sensing data (Meinel and Neubert 2004).OBIA correlates the segmentation parameters of images with spatial attributes and constructs a mean variance (Eq.( 1), Woodcock and Strahler 1) based on the segmentation objects to evaluate the merits of the scale parameters.The ESP2 tool (an eCognition plug-in) evolved on the mean variance, iteratively generated image objects at multiple scale levels in a bottom-up approach, and measured LV (local variance) as the values of mean variance in a 3 Â 3 moving window, while extending the calculation of local variance from one band to multiple bands (Eq.( 2)).
where LV K is the mean variance of different objects in band k, C ki is the luminance value of the i th pixel in the object in band k, n is the number of pixels in an object and m is the total number of objects in an image.
LV is the mean value of LV K over l bands.The dynamic change in LV reveals the effect of segmentation from one object level to another.To show the dramatic degree of change, the rate of change (ROC, Eq. ( 3)) is defined in ESP2.The thresholds in the rates of change indicate the scale levels at which the image can be segmented in the most appropriate manner relative to the data properties at the scene level.
where L ¼ LV at the target level, and L À 1 ¼ LV at the next lower level.ROC À LV measures the amount of change in LV from one object level to another, the local maximum in the ROC À LV graph indicate the object levels at which the image can be segmented in the most appropriate manner.The ESP2 tool is programmed in Cognition Network Language (CNL) in the eCognition Developer 64 software (Tiede and Hoffmann 2006), and it is implemented as a customized process to be applied easily similar to other processes in object-based rule set creation in the Definiens software.Six user-defined parameters are adjustable: (1) step size of increasing scale parameter, (2) starting scale parameter for the analysis, (3) the use of an object hierarchy during segmentation, (4) number of loops (i.e.number of scales to be tested), (5) shape weighting and (6) compactness weighting.Parameters ( 5) and ( 6) are used as implemented in multiresolution segmentation (Benz et al.2004;Zhu et al. 2020).
The simple yet robust ESP2 tool enables fast and objective parametrization when performing multiresolution segmentation.

Correlation-based feature selection technique
In contrast to other dimensionality reduction techniques, such as those based on projection (e.g.principal component analysis) or compression, feature selection techniques do not alter the original representation of the variables but merely select a subset of them.Thus, they preserve the original semantics of the variables, thereby providing a visual characterization of the features (Espindola et al. 2006).
The CFS assessed the worth of a set of features using a heuristic evaluation function based on the correlation of features, and Hall and Holmes (2003) claimed that a superior subset of features should be correlated with classes highly uncorrelated to each other.Thus, the criteria for a feature subset containing k features can be evaluated using the following formula (Eq.( 4)).merit ¼ where f indicates the feature, c is the class, r is the Pearson correlation coefficient calculated by Eq. ( 5) (Hall 2000), r cf denotes the mean feature correlation with classes, r ff indicates the average feature intercorrelation and k denotes the number of attributes in the subset.In addition, the forward search was used to explore the feature space, and the search stop criterion was set when the heuristic takes three consecutive declines to avoid searching the entire feature space.In this study, Numerical Python was used to implement this feature selection algorithm.

Maximum likelihood classification
ML classification is a widely used supervised image classification method.It constructs classification functions by assuming the distribution of remote sensing data to a multidimensional normal Eq. ( 6) (Lu et al. 2014).
where pðXjh i Þ is the probability that a pixel X belongs to a class h i :K is the number of bands, then X ¼ ½x 1 ; x 2 ; . . .; x k , R i and l i are the covariance matrix and mean value of class h i , respectively (Shi and Xue 2017).
The ML classification method calculates the distribution function of each class by extracting the mean and variance characteristics of the region of interest (ROI) and then estimating the probability of a pixel belonging to a particular class.Due to its ability to handle non-symmetric class distributions in the spectral domain, the ML method usually achieves high accuracy when the training data have high resolution (Richards 2006;Yang et al. 2011;Tian et al. 2013).In this study, we employed ENVI software to perform ML classification.

Process of identification of effective gerbil holes
This study proposes a fast and automatic system for identifying effective gerbil holes using UAV imagery.The system consists of three main steps, as illustrated in Figure 2: 1. Preprocessing is performed on all samples to ensure the relative consistency of the initial features of all samples, and the spatial distribution of the desert grassland features is obtained.2. Training is performed on samples to obtain a set of thresholds or parameters that define the meaningful visual characteristics of effective gerbil holes.3. The validity and accuracy of the method are tested and evaluated on the test samples.
To compare the classification results, typical pixel-based ML classification is used to classify the UAV imagery.The details of the system are described in Figure 2.

Scale parameters
The scale parameter is the most critical segmentation parameter of multiresolution segmentation, as it directly affects the segmentation of the entire image.Figure 3 shows the changes in LV and ROC with increasing scale parameters.As the scale parameter grew from a single pixel (starting scale is 1) in steps of 2, LV gradually increased until it matched the objects in UAV imagery.Thereafter, as the scale parameter continued to  increase, the boundary information of the object was retained at a higher segmentation level, while the LV value remained at the level indicating a meaningful ground object.
ROC followed an opposite trend to LV.This trend revealed how fast the change in LV was in the transition from pixels to the smallest characteristics in the image.In the scale range of 20-40 intervals, the ROC decreased abruptly, indicating that there were no objects of interest with scales less than 40.The local peaks in the ROC curve theoretically represented the emergence of larger homogeneous objects as the scale increased and indicated the appropriate parameters for image segmentation.However, the variation induced by the background segmentation also generated peaks, making the graph interpretation complicated.We selected the most dominant peaks that were clearly separated from their neighborhood as indicators for optimal scale parameters, namely, 45, 93, 107, 153 and 173.
The segmentation results are shown in Figure 4, displaying significant differences between the different scale parameters.When the scale parameter was set to a small value (e.g.45, 93), the characteristic difference of the target object (gerbil holes) was strongly influenced.Conversely, when the scale parameter was set too large (e.g.153, 173), the image segmentation was relatively coarse, and the gerbil hole boundaries were low.Additionally, other grassland ground objects around the gerbil holes were incorporated into the target object.After continuous adjustment in the training set, we found an optimal scale parameter of 107 that showed a clear separation between bare soils, grassland and gerbil holes.

Shape factors and compactness values
The shape factor represents the weight of the shape in homogeneity, and the sum of its value with that of the color factor is 1.0, determining the boundary fit for the image object.When optimizing the shape factor, iterative adjustments were made on the training set with the optimal scale parameter of 107.We found that the degree of coincidence of the gerbil hole boundary first increased and then decreased with increasing shape parameter.A small shape factor allowed for a greater influence of the color feature on the segmentation results, while a larger shape factor reduced the spectral homogeneity inside the gerbil holes, causing some of the gerbil holes to be incorrectly grouped with the bare soil.To further refine the object generation, compactness and smoothness values were calibrated.Lower weights on compactness (e.g.0.1 or 0.2) increased the smooth factor and segmented more objects with linear shapes, which facilitated the extraction of grass patches.Conversely, a higher compactness parameter (e.g.0.5 or 0.7) increased the compactness of the object and decreased the smoothness, leading to a poor fit of the effective gerbil hole edges.After repeated adjustments and visual assessments on the training set, we found that a shape factor of 0.2 and a compactness value of 0.3 resulted in an obvious separation between bare soil, grass and effective gerbil holes (as shown in Figures 5  and 6).

Object-based classification result
To identify gerbil holes, the information of features is summarized and generalized to obtain feature thresholds for different ground objects, which is a key step in rule-based classification in OBIA.However, the large number of derived features per segmented object can lead to a time-consuming and subjective process of optimizing the feature subset (Ho and Rao 2016).The CFS method provides an algorithmic way to optimize feature subsets, and the visualization tool can make predictions about their thresholds of representative features.

Feature selection result
In the present study, 35 features were calculated for each object, including spectral, texture and geometry features.As the vegetation index was not defined in eCognition Developer, we created the excess green value (EXG, Eq. ( 7)) (Toomey et al. 2015;Ran et al. 2016) to characterize the spectral features of vegetation.The details of all features are given in Table 2.After the univariate filtering method in feature preprocessing, two features, length_thic and thickness, were eliminated, leaving a final number of 33 valid features.The feature subsets merit obtained during traversal according to Eqs. ( 5) and ( 6) are shown in Table 3.Although the available feature space for each object reached 35, there is a large amount of redundant information among them, and some are not related to the target class.For example, the correlations between max_diff and mean_r, mean_g and mean_b are as high as 0.93, 0.94 and 0.74, respectively.The correlation between  roundness and elliptic_f, density, and rectangularity is greater than 0.9.With the extended feature space, OBIA may increase the complexity of the classification and the demand for computing power (Li et al. 2016).According to the CFS algorithm (Algorithm 1), the final optimal feature subset was determined, which contains the EXG, gray-level co-occurrence matrix (GLCM) and roundness.
where R, G and B represent the pixel values of the red, green and blue bands, respectively.

Threshold prediction and classification
We created a scatter plot of the optimal feature subset to avoid subjectivity and randomness in feature threshold prediction, as shown in Figure 7. From the scatter plot, we found that the spectral characteristics of the three ground objects are highly different.
Grassland patches had the highest EXG value, exceeding 5, while soil patches had an EXG value in the 0-10 range.Gerbil holes had the lowest EXG value, less than 0. In terms of texture characteristics, the GLCM of most grassland patches was higher than that of grassland and bare soil, reaching 0.150.The GLCM of gerbil holes was in the 0.120-0.150range, while soil patches had the lowest GLCM characteristic of less than 0.120.The roundness of gerbil holes was lower than that of grassland and bare soil.Grassland and bare soil had a heavy overlap in this feature.As our main task was identifying gerbil holes, the shape remained a candidate feature value.The next step was to fine-tune the feature threshold based on prediction to obtain the rule set for UAV image classification.
Based on the top-down classification rule, we used the minimum value as the extraction threshold of grassland after summarizing the EXG of all vegetation objects.After repeated adjustments on the training set, some bare soil was misclassified as grassland when the EXG threshold was less than 5, whereas some grassland could not be identified when it was greater than 5. Therefore, we achieved an optimized EXG threshold of 5, which effectively distinguished non-vegetation and vegetation.We then summarized the EXG of effective gerbil holes and found that the threshold value of the EXG rule had a remarkable difference in identifying the accuracy of gerbil holes.After several attempts, we were able to largely separate effective gerbil holes from the background (bare soil) when the EXG threshold was set to À 5.However, other bare soils were also classified as targets, as most of them had spectral characteristics similar to those of the effective gerbil holes (Figure 8(a)).The reason for this could be differences in the spectral characteristics of UAV imagery due to vegetation species.Therefore, we concluded that extracting effective gerbil holes via additional features was necessary.
The GLCM is another critical feature that performs classification in desert grasslands.As shown in Figure 8(a), the process of UAV imagery classification still shows partial misclassification or leakage based on the EXG feature.Classifying ground objects further in desert grassland is necessary to improve accuracy and relies on texture features.The analysis of the UAV imagery properties indicates that the GLCM value of gerbil holes is in the range of 0.120-0.145,the GLCM value of grassland exceeds 0.151 and the GLCM value of bare soil is the lowest, less than 0.118.Therefore, non-vegetation objects can be separated from vegetation to a large extent by setting the GLCM to 0.151.In non-vegetation objects, effective gerbil holes can be extracted from the background by setting the GLCM in the range of 0.120-0.145.Figure 8(b) shows a spatial distribution of vegetation similar to the original image.However, we found that a minimal number of gerbil holes are still unidentified when the gerbil holes are extracted based on the texture feature.Some patches are mistakenly identified as gerbil holes.This phenomenon may be due to the overlapping of these patches with the texture characteristics of gerbil holes.Therefore, we need to identify gerbil holes further using shape features to eliminate such phenomena.
The effective gerbil holes are mostly subcircular, unlike the irregular shapes of most of the grassland and bare soil patches.This finding is consistent with the scatter plot analysis.The roundness property describes how similar an image object is to an ellipse, with a value equal to the radius of the largest closed ellipse minus the radius of the smallest closed ellipse.The closer an image object's shape is to an ellipsoid, the lower its circularity is.Effective gerbil holes can be extracted from UAV images using the roundness feature.However, distinguishing patches consisting of grassland and bare soil is difficult.While the main task of the study is to identify effective gerbil holes, the calculation of vegetation cover is not the primary focus.Nevertheless, roundness can still be used for gerbil hole identification, with iterative testing performed on the training set all effective gerbil holes can be extracted when the roundness feature is less than 0.9.A minimal number of patches with shapes similar to mouse holes were also extracted (Figure 8(c)).
The classification accuracy was gradually improved through different segmentation parameters, layer superposition and optimization splicing of classification rules.The final classification result is shown in Figure 8(d), with all recommended segmentation parameters and classification thresholds detailed in Table 4.

Maximum likelihood classification result
The ML classification results (Figure 9(a)) exhibit a large number of small spots, indicating the need for post-classification methods to improve the accuracy.Three methods were chosen for this purpose.First, spots smaller than 2 pixels were analyzed and sieved in Figure 9(b), but this had little effect on the classification results.In Figure 9(c), a minimum sieving size of 200 (the minimum area of gerbil hole patches) was used, which removed the noise on bare soil but also produced unclassified pixel spots.The majority/minority method was used in Figure 9(d), where the category of the central pixel was replaced by the category of majority/minority pixels in a kernel of size 3.This method removed some small spots.In Figure 9(e), the clump process was used, which used erosion and dilation to merge small patches.With a kernel value of 3, the majority of small spots could be merged into the background.However, increasing the kernel value to 4 in Figrue 9(f) resulted in a smoother image but also fragmented the effective gerbil hole patches.After analyzing the results, we used the clump method with a kernel value of 3 to remove the small spots from the ML classification results.

Accuracy assessment of the two classification methods
The two methods obtained similar grassland and soil patterns and could accurately locate the effective gerbil holes (Figure 10(c,d)).The small patch removal and fusion growth in the object-based method further improved the classification accuracy (Figure 10(e)).The grassland area produced by the rule-based method was smaller than that produced by ML, while the soil area was larger.ML measures the probability that a pixel belongs to a category and assigns the pixel to the most likely category, thereby easily disregarding the integrity of ground objects and inevitably resulting in fragmentation.The post classification-processing can effectively remove small spots in the classification results (Figure 10(f)).
For further specific analysis, we formed a matrix for accuracy assessment based on the validation points generated from test samples.Accuracy indicators, including producer's accuracy (PA), user's accuracy (UA), kappa coefficients and overall accuracy (OA) for all three categories, were calculated.Additionally, the number of validation points was strictly controlled because their redundancy or inadequacy affects the evaluation accuracy (Pal and Foody et al. 2010).Tortora's theory was used to determine the number of validation points (Tortora et al. 1978) using Eq. ( 8).
where B is the upper ða=kÞ Â 100 th percentile of the v 2 distribution with one degree of freedom.Let P i , i ¼ 1, 2, . . ., k, be the proportion of the pixel in the category, and let b i ¼ ½BP i ð1 À P i Þ=n 1=2 : There was no prior knowledge about the values of P i , and a 'worst' case calculation of sample size could be made assuming some P i ¼ 1=2 and b i ¼ b for all i.Here, we required an absolute precision of 65% for each proportion and a confidence coefficient of 95%.In the above equation, b i ¼ 0:05, i ¼ 1, 2, 3, and a ¼ 0:05: We obtained a random sample size of 2550 in the test data using ArcGIS10.5.The attribute class of each point was determined by visual interpretation to form a matrix for accuracy assessment.A total of 1241 of the 2550 validation points were in grassland, 1053 were in bare soil and 256 were in gerbil holes.
The final accuracy evaluation matrix is shown in Table 5.As a supervised classification, ML classification obtained an overall accuracy of 86.67% and a kappa coefficient of 0.77.Due to the integration of spectral, texture and geometric features, the overall accuracy of the object-based method reached 88.25%, which is 2.58% higher than that of ML.More importantly, the object-based method has a higher user accuracy than the ML method in identifying effective gerbil holes.The reason for this phenomenon may be because of the 'salt and pepper' noise generated by the ML method.Although clump classes were performed after classification, the high image resolution still showed a great amount of pepper noise in the results, reducing the classification accuracy.
Because the accuracy of the validation points method is likely to be unstable due to mixed objects (Ma et al. 2017a), the area-based accuracy assessment method was also performed in our research.We statistically analyzed the differences between the two classification methods in extracting the area of effective gerbil holes.The total number of effective gerbil holes for the test samples was 20, covering a total area of 725cm 2 .The number of effective gerbil holes identified by rule sets was 17, achieving a recall rate of 85%.The covered area of the object-based method was 604 cm 2 , with an accuracy of 83.31% in identifying the patch area.The number of effective gerbil holes identified by ML classification was 102, far exceeding the ground truth.However, the covered area of this method was only 570 cm 2 , with an accuracy of 78.63% in identifying the patch area.

Relationship between effective gerbil holes and grass coverage
To examine the relationship between effective gerbil holes and grass coverage, we created a 5 Â 4 grid using the fishnet tool in ArcGIS 10.5 (Figure 11(a)).The results showed a significant correlation between the area of effective gerbil holes and grass coverage.When grass coverage was less than 5%, there were few effective gerbil hole patches, which may be related to insufficient potential food resources.As grass coverage increased above 5%, the area of gerbil holes expanded with sparser grass coverage.The maximum area of effective gerbil holes was observed when grass coverage was in the range of 20-35%.However, as grass coverage continued to increase, the area of gerbil holes decreased, which may be because dense grasslands would undermine the escape ability of gerbils, increasing their risk of predation.Additionally, the mean area of gerbil hole patches decreased linearly with increasing grass coverage (Figure 11(b)).This phenomenon may be related to the increased grass cover, biomass and adequate food resources.In conclusion, the distribution of gerbil hole patches is influenced by grass coverage, and gerbils will only aggregate when the grassland is degraded to a certain degree.

Efficient classification methods for UAV imagery
The supervised classification methods in OBIA are extremely labor-intensive and timeconsuming, while the unsupervised classification methods are highly automated but have limited classification accuracy.In this article, an automatic method for the fast and accurate identification of effective gerbil holes in desert grasslands was proposed.Such approaches are crucial for carrying out near real-time rodent censuses.The key to this approach is defining the meaningful visual characteristics of effective gerbil holes.To overcome the image processing difficulties at high spatial resolution, we propose a data driven framework for parameter objectification rather than 'trial and error'.We gathered UAV imagery from five sample plots and applied ESP2 and CFS to infer segmentation parameters and classification thresholds.Based on our analysis, we recommend the following rule sets: scale parameter of 107, shape factor of 0.2, compactness value of 0.3, EXG feature less than À5, GLCM value in the 0.120-0.145range, and roundness feature value less than 0.9.This study compared two methods for identifying effective gerbil holes.Both methods achieved high classification accuracy, exceeding 85%.However, the OBIA method had higher classification accuracy than ML, indicating that texture and geometric features can provide extra details and improve separability among objects.Furthermore, the OBIA method demonstrated better performance in recognizing the shape integrity of each gerbil hole patch.The OBIA method outperformed in extracting the number and area of effective gerbil holes, achieving 84.10% and 83.37%, respectively.When applied to the specific task of rodent censuses, the proposed system demonstrated that detecting gerbil holes in desert grassland is feasible when employing a simple RGB camera mounted on a UAV.
OBIA is a promising approach not only for supervised classification but also for unsupervised classification.The accuracy of the rule-based unsupervised classification method is facilitated by the application of ESP2 and CFS techniques.Despite iterative adjustments in optimizing segmentation parameters and selecting features, the above results show that homogeneous segmentation provides more available features that are crucial for extracting objects with specific shapes and distinct texture features.Conversely, feature selection reduces the dimension of features and improves the efficiency of object- oriented classification.Using this research as an example, the process of configuring features for every object during data labeling took over 12 min.After our parameter optimization, it only takes 1.7 min to classify a sample, which is an 86% improvement in efficiency.Considering that rule-based OBIA does not require manual sample selection, only a statistical analysis to determine segmentation parameters and thresholds is applied to a large amount of data from the same batch of flights.The labor and time costs spent on classification are much less than those of supervised classification methods.

Influences of grass coverage on M. unguiculatus
The number of effective gerbil holes is the most important indicator for densities of M. unguiculatus, and the patches formed by gerbils are an important parameter of grassland desertification.This article, for the first time, reported the effects of grass coverage on M. unguiculatus at a low-altitude remote sensing scale.The number of M. unguiculatus varied unimodally with grass coverage.When grass coverage increases or decreases from a certain threshold, the number of M. unguiculatus declines.This interaction reveals that when grassland degrades to a certain extent, the number of M. unguiculatus increases.
Previous studies have shown that rodent control alone has low efficiency in alleviating grassland degradation.The M. unguiculatus population can still recover after the cessation of rodent control (Tang et al. 2019).Increasing grass coverage may be a more feasible and ecologically sound pathway for controlling the M. unguiculatus population in practice.

Conclusion
Object-based image analysis offers a promising framework to address spectral mixing of different classes at subpixel scales.Previous studies have revealed the improvement in classification accuracy with object-based compared to pixel-based methods up to 31%.However, these comparisons were mostly conducted based on supervised classification, which varied not only in data and methods but also in the quantity and quality of training samples.In this article, rule-based classification is compared with ML for the first time, and the results show that by incorporating specific features of the target class, the rule-based algorithm can achieve superior results compared to pixel-based supervised classification.Moreover, rule-based algorithm has the feature of unsupervised classification, which has wide application value for massive remote sensing data processing.
The ESP2 and CFS algorithms are critical for determining the universal rule set.ESP2 defines a series of meaningful segmentation scale parameters for the image.But it is important that these scale parameters not be straightforwardly linked to a certain object, which produces some trouble to find an appropriate value of scale parameter without performing some attempts.In the future research, we will try scale parameter estimation which be linked to target object, based on labeled data.In this study, only three features were selected based on the CFS algorithm, which limited the extraction accuracy of effective mouse holes.The number of features can be increased in future research to achieve a balance between image processing speed and classification accuracy The effectiveness of the rule-based classification method in extracting effective gerbil holes is affected by the timing of data acquisition and image resolution.During the spring when vegetation cover is low, gerbil hole patches are easily identified from UAV imagery.As vegetation coverage increases, occlusion and shadows in images become more prevalent, and the applicability of this method may be reduced.Our future studies will explore ways to better utilize multi-temporal data to improve segmentation and enhance the accuracy of effective gerbil hole identification.By incorporating texture and geometric features in the extraction of gerbil hole patches, the accuracy of this method is highly dependent on the ground sampling distance (GSD).Increasing the altitude of the UAV flight can improve the efficiency of image acquisition but may reduce image resolution and classification accuracy.Moreover, the thresholds of spectral features in the rule set may need to be adjusted when the UAV flight altitude changes.Further research is needed to validate these findings.

Disclosure statement
No potential conflict of interest was reported by the author.

Figure 1 .
Figure 1.Location of the study area on Ordos Plateau, China.

Figure 2 .
Figure 2. Process of gerbil hole identification and optimization based on UAV imagery.

Figure 3 .
Figure 3. ESP tool outputs for the UAV imagery Graphs depict changes in local variance (LV) (solid black) and rate of change (ROC) (solid blue) with increasing scale parameter.

Figure 4 .
Figure 4. Segmentation results of UAV imagery under different scale parameters.

Figure 5 .
Figure 5. Segmentation results of gerbil holes under different shape factors.

Figure 6 .
Figure 6.Segmentation results of gerbil hole under different compactness values.

Figure 7 .
Figure 7. Scatter plot of the distribution between optimal feature subset.

Figure 8 .
Figure 8. Classification results of UAV imagery based on spectral features, texture features and geometric features.

Figure 10 .
Figure 10.UAV imagery and partly enlarged detail of classification results.(a) UAV imagery of an entire sample plot, (b) Subset of UAV imagery, (c) OBIA result of subset, (d) ML classification result of subset, (e) OBIA with post classification and (f) MLC with post classification

Figure 11 .
Figure 11.Relationship between grass coverage and area of effective gerbil holes.(a) Area of gerbil holes and (b) Mean area.

Table 1 .
Sample plots distribution and the numbers of gerbil holes.
and Y represent two different feature vectors or a feature vector X and a target category Y.As shown in Algorithm 1, the CFS technique consists of feature preprocessing and feature selection.

Table 2 .
Image object information on brightness, spectral and geometric features.
GeometricArea, Width, border length, length/thickness, length/width, number of pixels, thickness, volume, width, asymmetry, border index, compactness, density, elliptic fit, roundness, shape index Geometric features refer to the geometric information of meaningful objects, which is crucial to the classification effect when the target object has a regular geometry.

Table 3 .
Correlation-based forward search process.

Table 5 .
Accuracy assessments of two methods.