Object-Based Land Cover Classification Using Integration of Lidar Data and Satellite Image

Object-Based Land Cover Classification Using Integration of Lidar Data and Satellite Image

A rewrited-in-english version from the original paper (check Acknowledgement)

July 29, 202316 min read

Abstract

The local government's need for geospatial information becomes important when green open space is difficult to find in urban areas. The information is a land cover map, which can be obtained from the classification process of high-resolution satellite images, but it still has limited spectral information where object classification also involves spatial characteristics and textures to get high accuracy value (Kushardono, 2017). These characteristics are minimal in passive sensor data but are mostly found in active sensor data (radar/LiDAR). By knowing the potential of the two sensors (satellite imagery and LiDAR), several researchers have conducted similar studies, including Awrangjeb et al., (2013), Uzar and Yastikli (2013), and Gilani et al., (2015). In this study, there are 3 main data, namely Pleiades images, nDSM, and intensity images. nDSM is derived data from the LiDAR elevation value, while the intensity image is formed from the object's reflection value at the NIR wavelength, each of which has been corrected and then interpolated into raster data and classified. The classification process is carried out using OBIA method with the Assign Class algorithm through a segmentation process. The resulting classification scheme produces 12 subclasses (within 4 main classes) from each sensor data classification to form a composition dataset. This integration process produces 3 compositional datasets, namely dataset A (Pleiades image only), dataset B (Pleiades-intensity) and dataset C (Pleiades-nDSM). With GIS analysis, an accuracy test was carried out and the accuracy value of dataset A was 44.44% and dataset B and C both produced an accuracy value of 63.89%. The accuracy value is very low when referring to Indonesian Land Use/Land Cover Classification National's Standard, because the number of 36 sample points is disproportionate (<20%) to the total number of objects that reach thousands.

Keywords: OBIA, land cover, Pleiades, LiDAR, nDSM, intensity, integration

1. Introduction

A land cover map can be obtained from the results of land classification using various methods, one of which is remote sensing techniques (Sampurno & Thoriq, 2016; Aristalindra et al, 2020; Prayogo & Basith, 2020). In a study conducted by Alonso and Malpica (2010), a study was conducted on the synergistic combination of sensors between LiDAR and multispectral imagery in mapping the distribution of buildings. From the research that has been done, the results obtained are that the Overall Accuracy classification using satellite imagery data and LiDAR is 97.12% while the Overall Accuracy classification using satellite imagery data without LiDAR data is 80.34%.

LiDAR (Light Detection And Ranging) is a remote sensing technology with active sensors in the form of lasers that are integrated with flying vehicles/aircraft or Unmaned Aerial Vehicles (UAVs) which are capable of scanning a large area (Istarno, 2016). While active sensor vehicles (radar/LiDAR) have advantages in terms of spatial (higher geometric accuracy with very minimum relief displacement), satellites have advantages in spectral terms (wider range of spectral channels/bands in identifying characteristics of objects) (Kushardono, 1997; Santosa, 2016; Aristalindra et al, 2020).

Theoretically also according to Kushardono (1997), the difference between the two types of sensors between passive sensors and active sensors provides different output but can provide the same interpretation of certain objects based on a reference or rule. Of course this can enrich the information or characteristics of an object so that it is assumed to increase the accuracy of the classification results. This study aims to explore the potential of integrated satellite imagery with LiDAR data in classifying land cover objects in urban areas that have heterogeneous land cover objects. The research results are expected to provide answers to the hypotheses that have been made and have reliability that can be measured through accuracy tests.

2. Materials and Methods

2.1 Data and Location Study

Study area coverage
Study area coverage

This research activity took place in Makassar City in 2013, precisely in Tamalanrea Indah Village. This location was chosen based on the researcher's consideration of the availability of data from two data provider agencies (PT. Asi Pudjiastuti Geosurvey and LAPAN Remote Sensing Information & Data Center) at adjacent locations and times. Some of the data used in this study are as follows:

1. LiDAR point cloud data (Location: Makassar City, Time of acquisition: 1 – 22 June 2012, Average point spacing: 0.47 meters, Data format: *las, Data source: PT. Asi Pudjiastuti Geosurvey)

2. Trajectory recording of LiDAR aircraft (Location: Makassar City, Time of acquisition: 1 – 22 June 2012, Flying height: 776 – 1038 meters above sea level, Data format: *trj, Data source: PT. Asi Pudjiastuti Geosurvey)

3. Pleiades satellite image data (Location: Makassar City, Time of acquisition: August 18, 2013, Spatial resolution: 0.5 meters, Data format: *TIFF, Data source: LAPAN Remote Sensing Information & Data Center)

2.2 Methodology

2.2.1 Point Cloud Data Correction & Filtering

The raw point cloud
The raw point cloud
The point cloud is divided into square blocks measuring 1008.0349 x 1008.0349 m2 into a total of twelve blocks. Next, a point cloud filtering process is carried out again from the default class to get as many as eight point cloud classes, namely the default class, ground, low vegetation, medium vegetation, high vegetation, building, low point and model keypoints.

2.2.2 normalized Digital Surface Model (nDSM)

nDSM from First Return pulse
nDSM from First Return pulse
nDSM from Last Return pulse
nDSM from Last Return pulse
In generating nDSM, imported point clouds are filtered based on their height relative to ground level (point clouds that are relatively lower than ground level are not used). Meanwhile, the point spacing values obtained from the stages of observing the characteristics of the point cloud which give an average value of 0.471651 meters are useful for determining the pixel size of the resulting raster. With this configuration, nDSM can be generated automatically and then exported in raster format.

2.2.3 Intensity Image

Intensity image
Intensity image
Intensity images with maintained radiometric quality can be obtained using image enhancement techniques, namely contrast stretching. Intensity images that have improved their radiometric quality by contrast stretching can then be further processed for classification.

2.2.4 Pleiades Satellite Image

Pleiades satellite image
Pleiades satellite image

2.2.5 Object-Based Image Analysis (OBIA) Classification

In the segmentation process, the values of the input parameters (scale, shape, and compactness) are determined by trial and error, that is, values are entered repeatedly to obtain a visually proportional segment shape. The input value also pays attention to the appearance of the segment/object in the image.

Each segmented sensor data is then classified one by one. The classification process used is by using the Assign Class algorithm, which is a simple classification algorithm in which each object is classified based on the criteria value with a threshold condition into a certain class. The Assign Class algorithm is unsupervised, where object classification does not use the training area as a reference for classification. Objects in the form of segments will be classified if the criteria values on the object meet the given threshold. These criteria are in the form of spectral criteria (layer value), spatial criteria (shape and area) and texture criteria.

2.2.6 Data Integration

After the results of the three classifications of sensor data are exported, it can be overlaid with the identify method to carry out a matching up analysis of object classes to form compositional datasets. The analysis is carried out by calculating the percentage of the number of objects that have the same matching class of all objects. The technique used to find entry objects that meet these conditions is by making SQL queries. The number of objects selected with the query is then calculated based on the total number of identified objects to get the percentage value of matched up object class, while the composition dataset that is formed can be processed into the accuracy testing stage. The calculation results that have been obtained can then be analyzed further.

2.2.7 Accuracy Test

The accuracy test process was carried out on the classification results of dataset A (Pleiades images), dataset B (Pleiades images with intensity images) and dataset C (Pleiades images with nDSM). The accuracy test is measured from the matched up of the classified land cover class with the ground truth of the Google Earth temporal image which is visually interpreted. The selected ground truth coordinate points are displayed in the form of ground check points, which are taken randomly and evenly throughout the land cover of the case study area with a total of 36 points.

3. Results and Discussions

3.1 Segmentation Results

3.1.1 Pleiades Image

Pleiades image segmentation
Pleiades image segmentation
The value of the input segmentation parameter, namely the shape of 0.8 and the scale of 50, indicates that the appearance of objects in the Pleiades image is very heterogeneous spatially, while the value of the parameter compactness of 0.4 indicates that the image has a normal texture that tends to be smooth to be able to distinguish the shape of one object from another.

3.1.2 nDSM

nDSM segmentation
nDSM segmentation
Based on the segmentation results, it can be seen that the nDSM image is different from the Pleiades image. The most striking difference is the scale parameter value of 20, which means it is smaller than the value in the Pleiades image. This could be due to the appearance of objects in the nDSM less than the appearance of objects in the Pleiades imagery. Meanwhile, the shape parameter has a minimum value of 0.1, which means that the homogeneity of the pixels in nDSM is much easier to form into segments spectrally than spatially. The value of the compactness parameter of 0.6 indicates that texture in nDSM is quite influential in distinguishing the shape of one object from another.

3.1.3 Intensity Image

Intensity image segmentation
Intensity image segmentation
Based on the value of the input parameter, the intensity image has the appearance of objects that tend to be spectrally heterogeneous because the shape parameter is quite small, which is equal to 0.3. The scale parameter value of 80 is intended so that the segment size is proportional to the very wide image coverage. Meanwhile, the value of the compactness parameter of 1.0 indicates that the roughness level of the intensity image is very high in distinguishing one object from another.

3.2 Classification Results

3.2.1 Exported Result

Pleaides classfication result
Pleaides classfication result
nDSM classification result
nDSM classification result
Intensity image classification result
Intensity image classification result
In addition to the visual appearance, the classification results also have attribute data stored in the attribute table which contains information for each object, namely CLASS and SUB-CLASS attributes. This information can then be included in the analysis by overlaying the identify method to match the attribute data from the results of other sensor data classifications.

3.2.2 Query Analysis & Overlay Result

The overlay identify process focuses on overlaying spatial data (classified objects) from two different sensor data while maintaining the attribute data of each sensor data. The overlay identify result in the number of objects from each dataset being multiplied, which is mainly caused by the number and size of objects in each sensor data during the segmentation process. In order to be able to carry out a match up analysis for object classes, information on the number of objects that meet the conditions where the KELAS_1 attribute is the same as the KELAS_2 attribute is needed. Data search based on these conditions is done by using SQL queries.

3.3 Classification Scheme

Classification scheme from 3 different data
Classification scheme from 3 different data

3.3.1 Pleaides Image

Based on the classification scheme, the Pleiades image is very superior in terms of spectral which has four channel bands (red, green, blue and NIR) and band ratio (RVI, NDVI and NDWI) in identifying various kinds of objects. In addition, there are also subclass orchard from the class vegetation not found in subclasses from other sensor data, which in their identification use spatial criteria and texture criteria to distinguish these subclasses from other subclasses of the class vegetation.

3.3.2 nDSM

From the nDSM classification scheme, there is a difference with the classification scheme in the Pleiades imagery. This difference is very visible in the subclass road which is included in the class open area while in the class built up area the subclass building is divided into subclass residential buildings and subclass infrastructure buildings. This is because objects in the form of roads have relatively the same elevation as open area objects compared to built-up area objects. Meanwhile, the division of building subclasses into two subclasses can be done easily based on relative elevation values. In addition, another difference is also found in the absence of subclass orchard, due to limited identification features.

3.3.3 Intensity Image

In this classification scheme, it can be seen that there are many similarities in the class arrangement with the classification scheme used in the Pleiades imagery. The most striking difference is in the subclass lawn which is included in the class open area instead of the class vegetation. This is because the object has intensity characteristics that tend to resemble open area and does not give a significant value to the first of many return image like other vegetation subclasses. In addition, another difference is also found in the absence of subclass orchard, due to limited identification features.

3.4 Accuracy Test Result

Among the 36 test points, in the dataset A, the number of objects that match the ground truth is 16 points resulting in an overall accuracy value of 44.44%. Meanwhile, with the integration of LiDAR sensor data, dataset B (Pleiades-intensity) obtained 23 accurate test points resulting in an overall accuracy value of 63.89%, while dataset C (Pleiades-nDSM) also obtained 23 accurate test points resulting in an overall accuracy value of 63.89%.

3.4.1 Dataset A

Percentage of accuracy in Dataset A (Pleaides image only)
Percentage of accuracy in Dataset A (Pleaides image only)

In terms of producer accuracy, there are four subclasses that are classified accurately, namely bushes, pond, lake and river, while the subclasses for lawn, swamp, bare ground and other paved surfaces have the lowest accuracy where they are not classified correctly (indicated by the omission value). Meanwhile, in terms of user accuracy, orchards are the easiest subclass to identify, while the other classes tend to be difficult to distinguish from one another (indicated by commission values).

3.4.2 Dataset B

Percentage of accuracy in Dataset B (Pleaides image + Intensity image)
Percentage of accuracy in Dataset B (Pleaides image + Intensity image)

In terms of the accuracy of the map producer, there are five subclasses that are classified accurately, namely bushes, pond, lake, river and building, while swamp, bare ground and other paved surfaces are the most inaccurate subclasses (indicated by large omission values). Furthermore, in terms of map user accuracy, there are five subclasses that are the easiest to identify, namely bushes, lake, river, bare ground and building, while there are four subclasses that are still difficult to identify, namely orchard, swamp, road and other paved surfaces.

3.4.3 Dataset C

Percentage of accuracy in Dataset C (Pleaides image + nDSM)
Percentage of accuracy in Dataset C (Pleaides image + nDSM)

Dataset C has good mapping accuracy for six subclasses, namely bushes, trees, pond, lake, river and bare ground which can be classified accurately. However, for lawn, swamps and other paved surfaces are still the most inaccurate subclasses (indicated by the omission value). Meanwhile, in terms of map user accuracy, dataset C has almost the same accuracy as dataset B, only with slightly lower building accuracy. Thus, the accuracy of the classification results for dataset C is still better than dataset A and is equivalent to dataset B but only differs in the distribution of the subclasses.

3.5 Object Class Matching Up Analysis Results

Percentage of matched up objects between Dataset B and C
Percentage of matched up objects between Dataset B and C

Based on the average value, dataset C obtained the highest match up for class objects, namely 45.78% compared to dataset B which was 36.61%, where the average percentage value produced in dataset C was quite significant when compared to dataset B. Below are some examples of mismatched classification evident with various causes each from Dataset B and C.

3.5.1 Dataset B

Mistmatched of classification due to noises
Mistmatched of classification due to noises
Mistmatched of classification due to shadowed objects
Mistmatched of classification due to shadowed objects
Mistmatched of classification due to land cover changes
Mistmatched of classification due to land cover changes

3.5.2 Dataset C

Mistmatched of classification due to pulse interaction with water
Mistmatched of classification due to pulse interaction with water
Mistmatched of classification due to spectral character differences
Mistmatched of classification due to spectral character differences
Mistmatched of classification due to segmentation differences
Mistmatched of classification due to segmentation differences

4. Conclusion

Based on the results and discussion of the research obtained, the following conclusions are obtained:

1) The results of the Pleiades image classification (dataset A) using the OBIA method in the case study area as a whole produce a fairly low accuracy of 44.44%.

2) The results of the Pleiades image classification that have been added to LiDAR data (intensity images and nDSM) where each forms dataset B and dataset C, both of which are able to increase the accuracy of the classification results by 19.45% or nearly 20% compared to dataset A which only utilizes a single Pleiades sensor,

3) The low value of the accuracy of the classification results in this study when referring to SNI Land Cover/Land Use (minimum 70%) can be caused by the number of sample points which are only 36 sample points, very far from the number which is proportional to the total number of classified objects which reach thousands, which should cover at least 20% of the total number of classified objects to be able to provide significant test results,

4) Dataset B and dataset C each produce a different percentage of match up for object classes, where based on the lowest and highest values in dataset B produce object match up in the range of 30% to 41%, and dataset C produces object match up in the range of 31% to 56%. It can be inferred that object class matching up does not affect the increase in the accuracy of the classification results against ground truth.

5. Acknowledgement

The authors declare no any competing interest while writing this article. You can read the original version (written in Bahasa Indonesia) in this link.

6. Reference 

  • Achsan, A. C. (2017). Pemanfaatan Citra Landsat untuk Klasifikasi Tutupan Lahan Lanskap Perkotaan Kota Palu. E-Journal Arsitektur Lansekap, 59.

  • Alonso, M., & Malpica, J. (2010). Satellite Imagery Classification with Lidar Data. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science.

  • Aristalindra, F., Santosa, P. B., Diyono & Subaryono. (2020). Evaluasi Pemanfaatan Citra Tegak Satelit Resolusi Tinggi untuk Percepatan Pembuatan Peta Blok Pajak Bumi dan Bangunan (PBB-P2) secara Partisipatif di Desa Triharjo, Kabupaten Bantul, DIY. Journal of Geospatial Information Science and Engineering, vol 3 no 1. https://doi.org/10.22146/jgise.55788

  • Awrangjeb, M., Zhang, C., & Fraser, C. (2013). Automatic extraction of building roofs using LiDAR data and multispectral Imagery. ISPRS Jorunal of Photogrammetry and Remote Sensing, 83, 1-18.

  • Danoedoro, P. (2006). Versatile Land-use Information for Local Planning in Indonesia. Centre for Remote Sensing and Spatial Information Science (CRSSIS): School of Geography, Planning and Architecture. The University of Queensland.

  • Gilani, S., Awrangjeb, M., & Lu, G. (2015). Fusion of LiDAR data and multispectral imagery for effective building detection based on graph and connected component analysis. The International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, 40, 65.

  • Guo, Q. H., Kelly, M., Gong, P., & Liu, D. (2007). An object-based classification approach in mapping tree mortality using high spatial resolution imagery. GIScience dan Remote Sensing.

  • Istarno. (2016). Buku Ajar Penginderaan Jauh Sensor Aktif Airborne Laser Scanning/Lidar. Yogyakarta: Departemen Teknik Geodesi UGM.

  • Jia, Y. (2015). Object-based land cover classification with orthophoto and lidar data. Stockholm: School of Architecture and the Built Environment (KTH) Stockholm, Sweden.

  • Kushardono, D. (1997). Model Klasifikasi Penutup Lahan Data Multisensor Radar-Optik. Majalah LAPAN no.83.

  • Kushardono, D. (2017). Klasifikasi Digital Penginderaan Jauh. Bogor: Penerbit IPB Press.

  • Mills, M. (2020, June 23). Elevation Grid Creation in Global Mapper: Creating a DTM. Retrieved from Insight from Blue Marble Geographics: https://blog.bluemarblegeo.com/2020/06/23/elevation-grid-creation-in-global-mapper-creating-a-dtm/

  • Prayogo, L. M. & Basith, A. (2020). Uji Performa Citra Worldview 3 dan Sentinel 2A untuk Pemetaan Kedalaman Laut Dangkal (Studi Kasus di Kepulauan Karimunjawa, Jawa Tengah). Journal of Geospatial Information Science and Engineering, vol 3 no 2. https://doi.org/10.22146/jgise.59572

  • Sampurno, R., & Thoriq, A. (2016). Klasifikasi Tutupan Lahan Menggunakan Citra Landsat 8 Operational Land Imager (OLI) di Kabupaten Sumedang. Jurnal Teknotan Vol.10.2.

  • Santosa, P. B. (2016). Evaluation of satellite image correction methods caused by differential terrain illumination. Forum Geografi, vol 30 no 1. doi: 10.23917/forgeo.v30i1.1768

  • SNI 7645. (2010). SNI 7645: 2010 tentang Klasifikasi Penutup Lahan. Badan Standardisasi Nasional.

  • Suradji, A. (2020, May 18). Webinar PT Waindo Specterra. Retrieved from Teknologi GIS Enterprise, Dashboard Operation, CSRT, airborne LiDAR dan Aplikasi Pemanfaatannya di Bidang Pertambangan: https://youtu.be/1e9LgCy1yqs

  • Uzar, M., & Yastikli, N. (2013). Automatic building extraction using LiDAR and aerial photographs. Boletim de Ciências Geodésicas, 153-171.

  • Wahdaniyat, H. (2019, August 20). Ruang Terbuka Hijau yang Masih Terpinggirkan di Indonesia. Retrieved from Direktorat Jenderal Cipta Karya, Kementerian Pekerjaan Umum dan Perumahan Rakyat: http://ciptakarya.pu.go.id/pbl/index.php/detail_berita/565/ruang-terbuka-hijau-yang-masih-terpinggirkan-di-indonesia

GISimage classificationLiDARRemote Sensingsatellite data

Join the community!

We're a place where geospatial professionals showcase their works and discover opportunities.

Explore More Articles