Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci., XL-7/W3, 569-573, 2015
© Author(s) 2015. This work is distributed under
the Creative Commons Attribution 3.0 License.
29 Apr 2015
Fusion of hyperspectral and lidar data based on dimension reduction and maximum likelihood
B. Abbasi1, H. Arefi1, B. Bigdeli1, M. Motagh1,2, and S. Roessner2 1Department of Geomatics and Surveying Eng., University of Tehran, Tehran, Iran
2GFZ German Research Centre for Geosciences, Section of Remote Sensing, 14473, Potsdam, Germany
Keywords: Lidar, Hyper-spectral, Fusion, Classification, Maximum Likelihood Abstract. Limitations and deficiencies of different remote sensing sensors in extraction of different objects caused fusion of data from different sensors to become more widespread for improving classification results. Using a variety of data which are provided from different sensors, increase the spatial and the spectral accuracy. Lidar (Light Detection and Ranging) data fused together with hyperspectral images (HSI) provide rich data for classification of the surface objects. Lidar data representing high quality geometric information plays a key role for segmentation and classification of elevated features such as buildings and trees. On the other hand, hyperspectral data containing high spectral resolution would support high distinction between the objects having different spectral information such as soil, water, and grass. This paper presents a fusion methodology on Lidar and hyperspectral data for improving classification accuracy in urban areas. In first step, we applied feature extraction strategies on each data separately. In this step, texture features based on GLCM (Grey Level Co-occurrence Matrix) from Lidar data and PCA (Principal Component Analysis) and MNF (Minimum Noise Fraction) based dimension reduction methods for HSI are generated. In second step, a Maximum Likelihood (ML) based classification method is applied on each feature spaces. Finally, a fusion method is applied to fuse the results of classification. A co-registered hyperspectral and Lidar data from University of Houston was utilized to examine the result of the proposed method. This data contains nine classes: Building, Tree, Grass, Soil, Water, Road, Parking, Tennis Court and Running Track. Experimental investigation proves the improvement of classification accuracy to 88%.
Conference paper (PDF, 1156 KB)

Citation: Abbasi, B., Arefi, H., Bigdeli, B., Motagh, M., and Roessner, S.: Fusion of hyperspectral and lidar data based on dimension reduction and maximum likelihood, Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci., XL-7/W3, 569-573,, 2015.

BibTeX EndNote Reference Manager XML