Integration of case-based reasoning and object-based image ...

2 downloads 0 Views 3MB Size Report
sensed data (Blaschke 2010; Duro, Franklin, and Dubé 2012; Gao 2008). There are two major stages in object-based image classification: image segmentation ...
GIScience & Remote Sensing, 2013 Vol. 50, No. 5, 574–589, http://dx.doi.org/10.1080/15481603.2013.842292

Integration of case-based reasoning and object-based image classification to classify SPOT images: a case study of aquaculture land use mapping in coastal areas of Guangdong province, China Yunyan Dua, Di Wua, Fuyuan Liangb* and Ce Lic a

State Key Laboratory of Resources and Environmental Information System, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Sciences, Beijing 100101, China; b Department of Geography, Western Illinois University, 1 University Circle, Macomb, IL 61455, USA; cThe Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences, Beijing 100094, China (Received 15 April 2013; accepted 4 September 2013) We present a method to integrate case-based reasoning (CBR) with object-oriented image classification to classify SPOT images. Images were first segmented into discrete objects at multiple scales. CBR was then used to classify these objects by comparing their geometric shapes, spectral characteristics, and textural measurements with those of the past objects prepared from archived SPOT images and land use data. Once enough past objects were accumulated, this method was able to successfully classify image objects with promising results as demonstrated by a case study of aquaculture land use mapping in coastal areas of Guangdong province, China. Keywords: object-based classification; case-based reasoning; aquaculture; land use mapping

1. Introduction Object-oriented image classification has recently become one of the most commonly used approaches in image classification (Blaschke 2010; Aplin and Smith 2011) and it generally outperforms the pixel-based classification in extracting thematic information from remotely sensed data (Blaschke 2010; Duro, Franklin, and Dubé 2012; Gao 2008). There are two major stages in object-based image classification: image segmentation and object classification (Definiens® 2009). In the segmentation stage, image is first segmented into objects at different scales. One of the key issues is to find the optimal segmentation scale through trial-and-error tests, human evaluation, or quantitative studies (e.g., Benz et al. 2004; Drăguţ, Tiede, and Levick 2010; Frohn and Chaudhary 2008; Kim et al. 2009; Möller et al. 2007). This process is normally time-consuming and tedious as the optimal scale usually varies for different kinds of applications and images (Blaschke 2010; Myint et al. 2011). In the classification stage, thematic classes are assigned to the objects using the nearest neighbour or knowledge-based classifiers (Definiens® 2009). The nearest neighbour classification starts with selecting training data for each class. Spectral characteristics and textural measurements of the training data are used to determine the spatial clustered location of each specific class in a multiple-dimensional space. The classifier calculates the distance between each object and the individual clustered location. Each object is then *Corresponding author. Email: [email protected] © 2013 Taylor & Francis

GIScience & Remote Sensing

575

assigned to the closest cluster. This classifier has been used in many studies (e.g., Myint et al. 2011; Robertson and King 2011). The knowledge-based classifier uses rules to classify images (Lu and Weng 2007). Rules are normally extracted from spectral and textural features and other ancillary data. For example, Myint et al. (2011) developed different expert rules to delineate buildings, impervious surface, swimming pools, lakes, and ponds from a QuickBird image. Kass et al. (2011) developed rules based on different texture-based measurements of about half of the image objects and then used these rules to classify the remaining objects. Artificial intelligence methods have been increasingly used in object-based image classification. Classification and regression trees (CART) were used in object-based mapping of vegetation, urban, and rangeland (Laliberte et al. 2007; Thomas et al. 2003; Yu et al. 2006). Duro, Franklin, and Dubé (2012) compared the performance of decision tree, random forest, and support vector machine on pixel-based against object-based image classification of agricultural landscapes. Man-made objects, such as building, bridge, and highway, were also successfully recognized by Inglada (2007) using a supervised learning approach. More artificial intelligence methods are expected to be used in conjunction with object-based image classification (Blaschke 2010) as these methods can extract more reliable rules to improve classification accuracy. In this article, we present a method to integrate another artificial intelligence approach, the case-based reasoning (CBR), into object-based image classification. The CBR uses knowledge derived from past cases to solve new problems (Watson 1998). CBR has been used to predict weather change (Jones and Roydhourse 1994), combat rangeland pests (Branting et al. 1999), examine air pollution (Lekkas et al. 1994), and support urban planning (Yeh and Shi 1999) and soil mapping (Shi et al. 2004). It has been integrated with pixel-based classifiers to classify multi-temporal synthetic aperture radar (SAR) images (Li and Yeh 2004) and Radarsat images (Qian et al. 2007). To our knowledge, this research is the first study to integrate CBR into objectbased image classification. In this research, images were first segmented into objects at various scales. CBR was used to classify the objects produced at a coarse scale. If not successful, the images of unclassified objects were further segmented at a finer scale. CBR was then used to classify the resulting smaller objects. This hierarchical procedure continues until most image objects were successfully classified at a reasonable scale. 2. Study area, data, and classification scheme Our study area is located along the coast of Guangdong province, China (Figure 1). In total, 15 scenes of geometrically rectified SPOT-5 image were used in this study. Thirteen out of these 15 images were acquired for the eastern coastal areas. The other two images were acquired for two test areas (Figure 1), one in the southern (Jiangmen area, 220 km2) and the other in the western coastal area (Shatian, 107 km2). We also used the hydrographical data extracted from the 1:50,000 National Geomatics Database of China in this study. Aquaculture is one of the key economic sectors in Guangdong province. It contributes 19.3% of the agricultural GDP of Guangdong as of 2011. Provincial and local governments are eager to have timely information about offshore aquaculture land for land use planning. In this study, we developed a method to automatically extract information of aquaculture land use to meet this expectation. Instead of classifying all land use categories, in this study we mainly focused on those sharing similar spectral characteristics with aquaculture land, including pond, reservoir, and salt field. Water in river channels was first masked out from the images using the hydrographical data and thus was not included in this study. Classification scheme is developed following the Specifications of Satellite Remote Sensing for 1:50,000 Island and Coastal Zone Investigation (Project 908 of State Oceanic

576

Y. Du et al.

Figure 1. A map showing our study area. The parallelograms show footprints of the SPOT images used in this study. These images were acquired on (A) 3 December 2003, (B) 5 November 2004, (C) 10 October 2004, (D) 10 October 2004, (E) 16 October 2004, (F) 16 October 2004, (G) 15 October 2004, (H) 23 May 2004, (I) 11 December 2004, (J) 14 September 2004, (K) 1 January 2005, (L) 23 October 2003, (M) 23 October 2003, (N) 19 December 2006, (O) 9 October 2004.

Administration of China 2006). Aquaculture consists of areas with open water that is used to raise sea food products such as fish and shrimp. This type of rectangular-shaped land usually has a low earth bank. A pond is defined as man-made small water body with water throughout the entire year. Most ponds are built for livestock water and irrigation. These water bodies are normally small. A reservoir is a kind of irregular-shaped anthropogenic feature built in the lowland of mountainous areas for the purpose of irrigation. On remote sensing images, reservoirs appear as dark blue water bodies surrounded by vegetation growing in mountains or undulating hills. Salt field, mainly in rectangular shape, consists of large and flat lands that are used to produce salt. Archived land use data are already available for the inland 5-km buffer zone of the coastal line in our study area. This data set was produced by visual interpretation from the same set of SPOT-5 images by a research group in the Institute of Geographical and Natural Resources Research, Chinese Academy of Sciences. Overall classification accuracy of this land use data set is 93.4% (Du, Liang, and Sun 2012). 3. Methodology Essentially, CBR is a process of solving new problems based on solutions of similar past problems. There are four major cycles in CBR (Aamodt and Plaza 1994): retrieving similar past cases, reusing knowledge of similar past cases to solve new problems, revising the solutions, and retaining solved cases for future problem solving. In this study, images were first segmented into discrete objects at multiple scales. These objects

GIScience & Remote Sensing

577

were treated as cases in CBR. CBR was then used to build a library with past image objects and find solutions for new objects from the library. Detailed procedures (Figure 2) are explained below. 3.1. Image segmentation Images were segmented using the Fractal Net Evolution Approach in Definiens® (Baatz and Schäpe 2000). As mentioned in Section 2, the spectral characteristics of these four land use categories tend to be similar while their shapes are quite different. Thus weighting factors of the shape, compactness, smoothness, and colour parameters in Definiens® were set to 0.4, 0.5, 0.5, and 0.6 respectively, to emphasize the potential influence of the shape while deemphasize that of the colour. The images were initially segmented at scales of 120, 80, 60, 40, and 20. After comparing the segmentation results with archived land use data, we considered the scales of 80 and 40 to be appropriate for this study. Therefore, CBR was used to classify the image objects created at these two scales following the procedures in Figure 2, which are explained in the following section. 3.2. CBR in image classification In CBR, every case is described by attributes that can be used to distinguish one case from another (Holt 1999; Shi et al. 2004; Watson 1998). A case library was first built up to store past cases. The new cases were then compared with past cases in the case library to identify all past similar cases for each new case. Solutions of these past similar cases were then accordingly assigned to each specific new case. We have developed a generic framework of CBR applications in geographical problem solving (Du, Liang, and Sun 2012). This study generally followed this framework to find similar past image objects for a new object. A brief explanation is provided below. 3.2.1. Case representation In CBR, a case usually consists of two components: its “problem description” and “solution” (Du, Liang, and Sun 2012; Holt 1999; Shi et al. 2004). Specific to this

SPOT images

Objects classified Segmentation at a coarse scale

Objects (New cases)

CBR reasoning

Similarity assessment

Y N

Extraction Images of unclassified objects

Case library (past cases)

Similar past cases found?

Segmentation at next scale Objects (New cases)

Similarity assessment

Figure 2. A flow chart showing the procedures to segment images and classify image objects in this study.

578

Y. Du et al.

study, the “solution” of an image object is its land use type. The “problem description” refers to a set of attributes that are used to distinguish one image object from another, including their geometric shapes, spectral characteristics, and textural measurements. As argued by Du, Liang, and Sun (2012), land use is significantly affected by its geographic settings. In this study, we considered the perpendicular distance from an image object to the coastal line. Pond and reservoir are usually further away from the coast line than the aquaculture and salt field. Accordingly, an image object is represented by four components: geometric shape, spectral characteristics, textural measurements, and geographical settings. Totally, 34 variables (not reported in this article) were calculated to represent each object. Principle component analysis (PCA) was then used to retain the essential while remove duplicated information (Davis 2002) among these 34 variables. Variables that contribute most to the first three most significant principle components in PCA were selected (Table 1) and used to describe each individual image object: Casei ¼ fðS1i ; S2i Þ þ G1i þ ðT1i ; T2i ; . . . ; T4i Þ þ Di g þ Scalei þ LCi

(1)

Where i is the ith image object; Si, Gi, and Ti, are the spectral characteristics, geometric shape, and textural measurements of object i; Di is the perpendicular distance from the ith object to the coastal line; Scalei is the scale at which the image object was created; LCi is the land use category of the ith object. 3.2.2. Case library construction Objects created from the images of the eastern coast (A to M in Figure 1) were first overlaid atop the archived land use data. Only the objects with one dominant land use type occupying at least 90% of the area were imported into the case library (Table 2). Totally, there are 6977 objects in the case library and 5847 out of them are aquaculture (Table 3). There are 572, 406, and 152 objects of pond, reservoir, and salt field, respectively. 3.2.3. Similarity evaluation and CBR A similarity index (SI) between each new image object (i) and every past object (j) in the case library was calculated using the following equation: Table 1.

Attributes used to describe the segmented objects in this study.

No. Indicator Components 1 2

S1 S2

Spectral Spectral

3

G1

Geometric

4

T1

Textural

5 6 7 8

T2 T3 T4 D1

Textural Textural Textural Spatial

Description Average of all digital numbers within a specific object on band 4 Average of all digital numbers within a specific object on all 4 bands pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi Shape index (4 perimenter=area): smooth objects have a lower shape index The grey level co-occurrence matrix (GLCM) homogeneity (all direction) of band 3 GLCM contrast (all direction) of band 1 GLCM entropy (45°) of band 1 GLCM dissimilarity (all direction) of band 4 Nearest distance: perpendicular distance from the centroid of a land object to the costal line

GIScience & Remote Sensing Table 2.

579

Past cases with their attributes in the case library. Attributes

Case ID 1 2 3 4 … 6976 6977

S1

S2

G1

T1

T2

T3

T4

D1

Scale

LC*

67.11 54.66 60.00 41.84 … 19.72 68.58

93.31 86.71 83.98 64.35 … 44.05 105.4

1.688 1.600 1.263 1.624 … 1.520 1.922

0.2087 0.3435 0.2567 0.4082 … 0.5969 0.170

368.8 331.7 258.8 166.0 … 10.76 183.6

6.654 5.637 5.261 4.827 … 3.266 7.603

10.93 7.837 11.30 7.220 … 3.233 8.172

3664 3367 1666 348.4 … 0 1436

40 40 40 40 … 80 80

1 1 2 3 … 1 4

Note: *LC (land use category) with 1, 2, 3, and 4 denoting aquiculture, pond, reservoir, and salt field, respectively.

Table 3.

Number and total area of old cases per land use type in the case library. Land use types

Scale 80 40

Case

Aquaculture

Pond

Reservoir

Salt field

Number Area (km2) Number Area (km2)

1191 546.45 4656 657.48

112 16.43 460 22.98

132 21.96 274 22.27

29 8.21 123 14.47

Total 1464 593.05 5513 717.20

SIcaseði;jÞ ¼ ws  SS ðcaseði;jÞÞ þ wG  SGðcaseði;jÞÞ þ wT  ST ðcaseði;jÞÞ þ wD  SDðcaseði;jÞÞ (2) where Ss(case(i,j)), SG(case(i,j)), ST(case(i,j)), and SD(case(i,j)) are the similarity coefficients of the components of spectral characteristics, geometric shape, textural measurements, and geographical settings. wS, wG, wT, and wD are weights assigned to these four similarity coefficients, respectively. These weights were determined based on the total number of variables in each component. A weight of 0.5 was assigned to wT, 0.25 to ws, and 0.125 to both wG and wD as there are four and two variables in the texture and shape components, respectively, and only one in the other two components. The coefficient of each component was calculated by the following equation: Sðcaseði;jÞÞ ¼

Pn 1 1n

 ð1 

jAkcaseðiÞ  Akcaseð jÞ j Þ Range k

(3)

where n is the total number of attributes taken into account, Akcase(i) and Akcase(j) are the kth attribute of cases i and j, respectively, Rangek is the range of the kth attribute. Objects created at scale 80 were first classified (Figure 3). To a specific new object, we identified all past objects with a SI higher than a threshold value (0.95 in this study). If only one similar past object was found, its land use type was directly assigned to this new case. In the case that more than one past similar object was found, land use type of the past object created at the same scale as the new object was accepted. Otherwise, land use type of the past object with the highest SI was accepted as the solutions to this specific

580

Y. Du et al. Image segmentation

New case (case i)

Similarity assessment

Accept as past case

T F

Figure 3.

Unclassified Objects

SPOT images

Case base (old cases)

Similarity Coefficients

Case 1 Case 2 … Case n

SIcase(i,1) SIcase(i,2) ……… SIcase(i,n)

Accuracy assessment

Classified objects

No similar cases founded

> threshold

N Y

Assign land use type Similar cases to this new case founded

A flow chart showing image classification procedures using CBR.

new object. If no similar cases were found for a specific object in the case library, it was temporarily labelled as “unclassified.” The image of this unclassified object was then further segmented at scale 40 into smaller objects, which were then classified following the aforementioned procedures. If an object generated at this finer scale still cannot be classified, it was eventually labelled as “unclassified.” 3.3. Experimental design and accuracy assessment Three experiments were designed to evaluate the applications of CBR in classifying remote sensing image. We first conducted a 10-fold cross-validation to test past objects in the case library. In this experiment, the past objects were randomly partitioned into 10 subsamples of approximately equal size. Ten tests were then performed. In each test, nine out of these 10 subsamples were used as past cases. The remaining one subset of objects was used as new objects and they were then classified using the procedures outlined in Section 3.2.3. In the second experiment, we used all past objects in the case library to classify objects created from images of the two test areas (N and O in Figure 1). Maximum likelihood classification (MLC) was also employed to classify the images of the two test areas. Classification results of CBR and MLC were compared to evaluate the effectiveness of CBR in image classification. We prepared error matrixes to evaluate the classification accuracy. In this study, we compared the classification results against the archived land use data in terms of their areal extent (Congalton 1991; Foody 2002). For the purpose of accuracy assessment, we only retained the aquaculture, pond, reservoir, and salt field and all other categories in the archived land use data were merged and recoded as “others.” We then overlaid our classification results over the archived land use data to calculate the area that was correctly classified and misclassfied. 4. Results and discussion 4.1. Classification accuracy assessment The overall classification accuracy of the 10-fold cross-validation experiment ranges from 95.34% to 97.44% with an average of 96.46% (Table 4). Among the 10 tests, less than 0.2% of aquaculture lands was unclassified and only 1.16% was misclassified. Both

Test 3 Classification

Test 2 Classification

Test 1 Classification

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

128.96 1.07 0.10 1.07 0.09 98.29

118.17 0.63 0.28 0.50 0.41 98.82

122.32 0.72 0.21 1.06 0.03 98.40

Aquaculture

Error matrixes of the 10-fold cross-validation experiment.

Area (km2)

Table 4.

0.38 2.09 0.36 0.00 0.00 73.88

1.26 2.48 0.04 0.00 0.00 65.59

0.55 2.50 0.96 0.00 0.00 62.40

Pond

0.38 0.26 4.58 0.00 0.00 87.78

0.45 0.53 3.26 0.00 0.00 76.89

0.87 0.17 3.34 0.00 0.00 76.33

Reservoir

Reference

1.71 0.00 0.00 0.71 0.00 29.26

2.01 0.00 0.00 1.09 0.00 35.15

0.60 0.00 0.00 1.51 0.00 71.55

Salt field

(continued )

96.18

98.12 61.13 90.90 39.88

95.34

96.95 68.10 91.06 68.43

96.17

98.38 73.76 74.11 58.75

User’s accuracy (%)

GIScience & Remote Sensing 581

(Continued).

Test 6 Classification

Test 5 Classification

Test 4 Classification

Area (km2)

Table 4.

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

122.76 0.87 0.24 0.52 0.24 98.68

Ground truth

112.34 0.31 0.25 1.30 0.38 98.37

122.01 0.68 0.07 0.36 0.35 99.10

Aquaculture

0.75 2.77 0.28 0.00 0.00 73.02

0.61 2.89 0.28 0.00 0.00 76.36

0.70 3.04 0.11 0.00 0.00 79.02

Pond

0.83 0.31 3.40 0.00 0.00 74.93

0.11 0.32 3.45 0.00 0.00 88.84

0.70 0.49 3.26 0.00 0.00 73.21

Reservoir

Reference

0.39 0.00 0.00 1.67 0.00 80.92

0.39 0.00 0.00 1.07 0.00 73.26

0.55 0.00 0.00 1.07 0.00 66.21

Salt field

96.72

98.42 70.13 86.80 76.15

96.80

99.02 82.13 86.57 45.14

97.00

98.43 72.17 94.77 75.02

User’s accuracy (%)

582 Y. Du et al.

Test 10 Classification

Test 9 Classification

Test 8 Classification

Test 7 Classification

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

113.54 0.60 0.11 0.59 0.56 98.86

118.41 0.85 0.48 0.68 0.10 98.33

117.22 0.36 0.05 0.67 0.06 99.08

128.59 0.45 0.77 0.84 0.05 98.42

0.48 2.64 0.41 0.00 0.00 74.78

1.54 3.48 0.28 0.00 0.00 65.66

0.83 2.49 0.53 0.00 0.00 64.73

0.68 3.46 0.55 0.00 0.00 73.76

0.06 0.26 4.43 0.00 0.00 93.41

0.25 0.47 4.40 0.00 0.00 85.88

0.52 0.06 3.22 0.00 0.00 84.61

0.85 0.37 2.64 0.00 0.00 68.52

0.17 0.00 0.00 2.39 0.00 93.50

0.48 0.00 0.00 2.44 0.00 83.60

1.06 0.00 0.00 0.90 0.00 45.78

1.04 0.00 0.00 1.42 0.00 57.61

97.44

99.38 75.45 89.52 80.12

96.17

98.12 72.52 85.28 78.16

96.75

97.98 85.29 84.72 57.32

96.05

98.04 80.97 66.63 62.79

GIScience & Remote Sensing 583

584

Y. Du et al.

producer’s and user’s accuracies of aquaculture land are consistently over 96% with an average of 98.64% and 98.28%, respectively. By contrast, producer’s and user’s accuracies of other land use types vary significantly. For example, user’s accuracy of salt field ranges from 39.88% to 80.12% with an average of 64.18. Producer’s accuracy of pond ranges from 62.40% to 79.32% with an average of 70.88%. Figure 4 shows the classification results of the two test areas. The overall classification accuracy is 83.26% and 90.84% for Jiangmen and Shatian test areas, respectively (Tables 5 and 6). The highest producer’s and user’s accuracies, ranging from 54.93% to74.73%, were produced by aquaculture in both test areas. The user’s and producer’s accuracies of pond and reservoir are less than 3% in Jiangmen area and between 10.81% and 50.28% in Shatian area. Classification accuracy is positively related to the total number of past objects of each land use category in the case library. The aquaculture produced the highest producer’s and user’s accuracies in both two experiments (Tables 4–6) as it is the dominant type of the past objects in the case library (Table 3). For the categories with fewer past objects in the case library, their classification accuracies are significantly lower than that of the aquaculture. Therefore, a representative and complete case library is the key to the successful applications of CBR in image classification. This is also the reason why the classification accuracies of pond and reservoir are high in the 10-fold cross-validation experiment. For each new object, there are roughly nine past objects with the same type of land use available in the case library. It is easy for CBR to find similar past cases for the new case and the classification accuracy thus tends to be high. This is also true for the classification results of the two test areas. Aquaculture produced significantly higher classification accuracy than pond and reservoir as there are much more past objects in the case library for CBR.

Figure 4.

CBR classification results of the two test areas: (A) Jiangmen and (B) Shatian.

GIScience & Remote Sensing Table 5.

585

Confusion matrix of the CBR classification results of Jiangmen test area. Reference

Area (km2) Classification Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Table 6.

Aquaculture Pond Reservoir 30.05 0.19 0.00 0.64 9.49 74.43

0.34 0.02 0.00 0.00 0.49 2.77

0.72 0.09 0.00 0.00 0.11 0.01

Salt field 0 0 0 0 0

Others 23.59 0.55 0.42 2.76 149.80 84.58

User’s accuracy (%) 54.93 2.77 0.02 0.00 93.69 82.03

Confusion matrix of the CBR classification results of Shatian test area. Reference

Area (km2) Classification Aquaculture Pond Reservoir Salt field Others Producer’s accuracy (%) Overall accuracy (%)

Aquaculture Pond Reservoir 6.27 0.00 0.00 1.74 0.50 73.69

0.55 0.17 0.03 0.06 0.80 10.81

0.16 0.00 0.03 0.00 0.07 12.21

Salt field 0 0 0 0 0

Others 3.20 0.17 0.09 2.82 90.20 93.49

User’s accuracy (%) 61.55 50.28 22.20 0.00 98.50 90.46

4.2. Comparison with MLC classification The confusion matrixes (Tables 5–8) show that CBR outperforms the MLC classifier in classifying SPOT-5 images of the two test areas as it further considers spatial arrangements of pixels. In Jiangmen area, CBR produced a significantly higher overall classification accuracy (82.03%) than the MLC did (73.78%). The overall CBR classification accuracy of Shatian area (90.46%) is also higher than that of the MLC (87.08%). Regarding to individual category, CBR exceeds the MLC classifier in classifying aquaculture as shown by the higher producer’s and user’s accuracies of aquaculture. However, producer’s and user’s accuracies of pond and reservoir in CBR classification results are significantly lower than that of the MLC classification. This is because there are not enough past cases in the case library for these two categories. Another reason is probably because the past cases were derived from a different area. Objects generated from images of two different areas may have significantly different spectral characteristics, geometric shapes, and texture measurements. However, CBR shows huge potential in image classification if there are enough past cases derived from the same area in the case library, as demonstrated in the 10-fold cross-validation experiment.

586

Y. Du et al.

Table 7.

Confusion matrix of the MLC classification results of Jiangmen test area. Reference

Area (km2) Classification Aquaculture Pond Reservoir Others Producer’s accuracy (%) Overall accuracy

Table 8.

Aquaculture

Pond

Reservoir

23.48 9.26 1.52 6.11 58.17

0.19 0.36 0.16 0.15 42.02

0.03 0.18 0.63 0.09 68.16

User’s Others accuracy (%) 26.80 12.36 0.65 137.30 77.52

46.50 1.63 21.13 95.58 73.78

Confusion matrix of the MLC classification results of Shatian test area. Reference

Area (km2) Classification Aquaculture Pond Reservoir Others Producer’s accuracy (%) Overall accuracy

Aquaculture

Pond

5.31 1.50 1.33 0.37 62.44

0.34 0.61 0.31 0.36 37.54

User’s Reservoir Others accuracy (%) 0.00 0.02 0.15 0.09 57.04

0.73 8.23 0.54 86.99 90.16

83.20 5.86 6.57 99.07 87.08

4.3. Implication of object-oriented image classification and land cover mapping Both CBR in this study and the membership function classifier in Definiens® use rules to classify image objects. The membership function classifier uses rules that are derived from expert knowledge or training data. Expert-derived rules tend to be subjective. Their effectiveness in image classification is dependent upon how thorough the expert knows the difference of spectral characteristics, geometric shapes, and textual measurements among the different types of land use. Rules extracted from training data are subjected to the completeness and representativeness of the training data. This actually remains a challenge in image classification, particularly for study areas with complex and heterogeneous landscapes (Lu and Weng 2007). By contrast, CBR extracts knowledge from past cases. Such rules are objective as they are neither constrained by user’s expert knowledge nor the training data. The key issue of CBR applications in image classification is to accumulate enough number of past cases in the case library as indicated by the 10-fold cross-validation experiment in this study. To build such a complete and representative case library is actually no more a challenge as archived images and land use data sets are already widely available today. Objects can be routinely generated from archived images at various scales. Land use of these objects can be obtained from the archived data sets. For example, global land cover mapping has been performed using images acquired by the Advanced Very High Resolution Radiometer (AVHRR) (Hansen et al. 2000; Loveland et al. 2000), the Vegetation Instrument onboard SPOT satellites (Bartholomé and Belward 2005), the Moderate Resolution Imaging Spectroradiometer (MODIS) (Friedl et al. 2002). National

GIScience & Remote Sensing

587

land cover databases are also available for many countries. For example, the USGS Earth Resources Observation and Science (EROS) Center has developed National Land Cover Database of 1992, 2001, and 2006 for the conterminous United States (Fry et al. 2011; Homer et al. 2007; Vogelmann et al. 2001). Many other studies have presented land mapping results at regional or even finer scales from various sources of remote sensing images (e.g., Gao 2008; Myint et al. 2011). These data sets undoubtedly allow us to use CBR, in conjunction with object-based image classification, to update land use thematic maps from similar types of remote sensing imagery. 5. Conclusions Land use data derived from remotely sensed images are available at various scales. Integration of CBR with object-based image classification allows users to extract knowledge from archived land use data to classify new remote sensing images to produce or update thematic land use maps. As demonstrated in this study, the key to the successful applications of CBR in image classification is to build a complete and representative case library from archived land use products. It is actually no more a challenge given that so many multi-resolution land cover databases have already been produced. Furthermore, once new image objects are correctly classified, they can be absorbed into the case library and used as for future image classification.

Funding Funding of this study was provided by National Science Foundation of China [grant number 41071250]; the State Science and Technology Support Program of China [grant number 2011BAH23B04] the Innovation Projects of the State Key Laboratory of Resource and Environment Information System, Chinese Academy of Sciences [grant number: 088RA500KA].

References Aamodt, A., and E. Plaza. 1994. “Case-Based Reasoning: Foundational Issues, Methodological Variations, and System Approaches.” AI Communications 7: 39–59. Aplin, P., and G. M. Smith. 2011. “Introduction to Object-Based Landscape Analysis.” International Journal of Geographical Information Science 25: 869–875. Baatz, M., and M. Schäpe. 2000. “Multiresolution Segmentation - an Optimization Approach for High Quality Multi-Scale Image Segmentation.” In Angewandte Geographische InformationsVerarbeitung XII, edited by J. Strobl, T. Blaschke, G. Griesebner, 12–23. Heidelberg: Wichmann Verlag. Bartholomé, E., and A. S. Belward. 2005. “GLC2000: A New Approach to Global Land Cover Mapping From Earth Observation Data.” International Journal of Remote Sensing 26: 1959–1977. Benz, U. C., P. Hofmann, G. Willhauck, I. Lingenfelder, and M. Heynen. 2004. “Multi-Resolution, Object-Oriented Fuzzy Analysis of Remote Sensing Data for GIS-Ready Information.” ISPRS Journal of Photogrammetry and Remote Sensing 58: 239–258. Blaschke, T. 2010. “Object Based Image Analysis for Remote Sensing.” ISPRS Journal of Photogrammetry and Remote Sensing 65: 2–16. Branting, L. K., J. D. Hastings, and J. A. Lockwood. 1999. “Integrating Cases and Models for Predictions in Biological Systems.” AI Applications 11: 29–48. Congalton, R. G. 1991. “A Review of Assessing the Accuracy of Classifications of Remotely Sensed Data.” Remote Sensing of Environment 37: 35–46. Davis, J. C. 2002. Statistics and Data Analysis in Geology, 656. New York: John Wiley & Sons. Definiens®, A. G. 2009. Definiens Developer 7–User Guide. München: Definiens AG.

588

Y. Du et al.

Drăguţ, L., D. Tiede, and S. R. Levick. 2010. “ESP: A Tool to Estimate Scale Parameter for Multiresolution Image Segmentation of Remotely Sensed Data.” International Journal of Geographical Information Science 24: 859–871. Du, Y., F. Liang, and Y. Sun. 2012. “Integrating Spatial Relations Into Case-Based Reasoning to Solve Geographic Problems.” Knowledge-Based Systems 33: 111–123. Duro, D. C., S. E. Franklin, and M. G. Dubé. 2012. “A Comparison of Pixel-Based and ObjectBased Image Analysis with Selected Machine Learning Algorithms for the Classification of Agricultural Landscapes Using SPOT-5 HRG Imagery.” Remote Sensing of Environment 118: 259–272. Foody, G. M. 2002. “Status of Land Cover Classification Accuracy Assessment.” Remote Sensing of Environment 80: 185–201. Friedl, M. A., D. K. McIver, J. C. F. Hodges, X. Y. Zhang, D. Muchoney, A. H. Strahler, C. E. Woodcock, S. Gopal, A. Schneider, and A. Cooper. 2002. “Global Land Cover Mapping From MODIS: Algorithms and Early Results.” Remote Sensing of Environment 83: 287–302. Frohn, R. C., and N. Chaudhary. 2008. “Multi-Scale Image Segmentation and Object-Oriented Processing for Land Cover Classification.” GIScience & Remote Sensing 45: 377–391. Fry, J., G. Xian, S. Jin, J. Dewitz, C. Homer, L. Yang, C. Barnes, N. Herold, and J. Wickham. 2011. “Completion of the 2006 National Land Cover Database for the Conterminous United States.” Photogrammetric Engineering and Remote Sensing 77: 858–864. Gao, J. 2008. “Mapping of Land Degradation From ASTER Data: A Comparison of Object-Based and Pixel-Based Methods.” GIScience & Remote Sensing 45: 149–166. Hansen, M. C., R. S. DeFries, J. R. Townshend, and R. Sohlberg. 2000. “Global Land Cover Classification at 1 Km Spatial Resolution Using a Classification Tree Approach.” International Journal of Remote Sensing 21: 1331–1364. Holt, A. 1999. “Applying Case-Based Reasoning Techniques in GIS.” International Journal of Geographical Information Science 13: 9–25. Homer, C., J. Dewitz, J. Fry, M. Coan, N. Hossain, C. Larson, N. Herold, A. McKerrow, J. N. VanDriel, and J. Wickham. 2007. “Completion of the 2001 National Land Cover Database for the Counterminous United States.” Photogrammetric Engineering and Remote Sensing 73: 337–341. Inglada, J. 2007. “Automatic Recognition of Man-Made Objects in High Resolution Optical Remote Sensing Images by SVM Classification of Geometric Image Features.” ISPRS Journal of Photogrammetry and Remote Sensing 62: 236–248. Jones, E. K., and A. Roydhouse. 1994. “Intelligent Retrieval of Historical Meteorological Data.” Artificial Intelligence 8: 43–54. Kass, S., C. Notarnicola, and M. Zebisch. 2011. “Identification of Orchards and Vineyards with Different Texture-Based Measurements by Using an Object-Oriented Classification Approach.” International Journal of Geographical Information Science 25: 931–947. Kim, M., M. Madden, and T. A. Warner. 2009. “Forest Type Mapping Using Object-Specific Texture Measures From Multispectral Ikonos Imagery: Segmentation Quality and Image Classification Issues.” Photogrammetric Engineering and Remote Sensing 75: 819–829. Laliberte, A. S., E. L. Fredrickson, and A. Rango. 2007. “Combining Decision Trees with Hierarchical Object-Oriented Image Analysis for Mapping Arid Rangelands.” Photogrammetric Engineering and Remote Sensing 73: 197–207. Lekkas, G. P., N. M. Avouris, and L. G. Viras. 1994. “Case-Based Reasoning in Environmental Monitoring Applications.” Applied Artificial Intelligence 8: 359–376. Li, X., and A. G. Yeh. 2004. “Multitemporal SAR Images for Monitoring Cultivation Systems Using Case-Based Reasoning.” Remote Sensing of Environment 90: 524–534. Loveland, T. R., B. C. Reed, J. F. Brown, D. O. Ohlen, Z. Zhu, L. Yang, and J. W. Merchant. 2000. “Development of a Global Land Cover Characteristics Database and IGBP DISCover From 1 Km AVHRR Data.” International Journal of Remote Sensing 21: 1303–1330. Lu, D., and Q. Weng. 2007. “A Survey of Image Classification Methods and Techniques for Improving Classification Performance.” International Journal of Remote Sensing 28: 823–870. Myint, S. W., P. Gober, A. Brazel, S. Grossman-Clarke, and Q. Weng. 2011. “Per-Pixel Vs. ObjectBased Classification of Urban Land Cover Extraction Using High Spatial Resolution Imagery.” Remote Sensing of Environment 115: 1145–1161.

GIScience & Remote Sensing

589

Möller, M., L. Lymburner, and M. Volk. 2007. “The Comparison Index: A Tool for Assessing the Accuracy of Image Segmentation.” International Journal of Applied Earth Observation and Geoinformation 9: 311–321. Project 908 of State Oceanic Administration of China. 2006. Specifications of Satellite Remote Sensing for Island and Coastal Zone Investigation, 61. Beijing: China Ocean Press. Qian, J., X. Li, B. Ai, G.-O. Y. Anthony, X.-Y. Chen, K. Liu, and L. Zhuo. 2007. “Land-Use ShortTerm Change Detection Based on Time-Series-Case-Based-Reasoning (TSCBR).” Journal of Natural Resources 5: 735–746. Robertson, L. D., and D. J. King. 2011. “Comparison of Pixel-and Object-Based Classification in Land Cover Change Mapping.” International Journal of Remote Sensing 32: 1505–1529. Shi, X., A.-X. Zhu, J. Burt, F. Qi, and D. Simonson. 2004. “A Case-Based Reasoning Approach to Fuzzy Soil Mapping.” Soil Science Society of America Journal 68: 885–894. Thomas, N., C. Hendrix, and R. G. Congalton. 2003. “A Comparison of Urban Mapping Methods Using High-Resolution Digital Imagery.” Photogrammetric Engineering and Remote Sensing 69: 963–972. Vogelmann, J. E., S. M. Howard, L. Yang, C. R. Larson, B. K. Wylie, and N. Van Driel. 2001. “Completion of the 1990s National Land Cover Data Set for the Conterminous United States From Landsat Thematic Mapper Data and Ancillary Data Sources.” Photogrammetric Engineering and Remote Sensing 67: 650–662. Watson, I.. 1998. Applying Case-Based Reasoning: Techniques for Enterprise Systems. USA: Morgan Kaufmann Publishers Inc., San Francisco. Yeh, A. G. O., and X. Shi. 1999. “Applying Case-Based Reasoning to Urban Planning: A New Planning Support System Tool.” Environment and Planning B: Planning and Design 26: 101–116. Yu, Q., P. Gong, N. Clinton, G. Biging, M. Kelly, and D. Schirokauer. 2006. “Object-Based Detailed Vegetation Classification with Airborne High Spatial Resolution Remote Sensing Imagery.” Photogrammetric Engineering and Remote Sensing 72: 799–811.

Suggest Documents