Abstract
Current and forthcoming sensor technologies and space missions are providing remote sensing scientists and practitioners with an increasing wealth and variety of data modalities. They encompass multisensor, multiresolution, multiscale, multitemporal, multipolarization, and multifrequency imagery. While they represent remarkable opportunities for the applications, they pose important challenges to the development of mathematical methods aimed at fusing the information conveyed by the input multisource data. In this framework, the present chapter continues the discussion of remote sensing data fusion, which was opened in the previous chapter. Here, the focus is on data fusion for image classification purposes. Both methodological issues of feature extraction and supervised classification are addressed. On both respects, the focus is on hierarchical image models rooted in graph theory. First, multilevel feature extraction is addressed through the latest advances in Mathematical Morphology and attribute profile theory with respect to component trees and trees of shapes. Then, joint supervised classification of multisensor, multiscale, multiresolution, and multitemporal imagery is formulated through hierarchical Markov random fields on quad-trees. Examples of experimental results with data from current VHR optical and SAR missions are shown and analysed.
Original language | English |
---|---|
Title of host publication | Signals and Communication Technology |
Publisher | Springer |
Pages | 277-323 |
Number of pages | 47 |
DOIs | |
Publication status | Published - 2018 |
Publication series
Name | Signals and Communication Technology |
---|---|
ISSN (Print) | 1860-4862 |
ISSN (Electronic) | 1860-4870 |
Bibliographical note
Funding Information:This work was partly supported by the French Space Agency (Centre National d’Etudes Spatiales, CNES) through contract no. 8361. The authors would like to thank CNES, the Italian Space Agency (ASI), and GeoEye Inc. and Google Crisis Response for providing the Pléiades, COSMO-SkyMed, and GeoEye imagery used for experiments.
Funding Information:
Acknowledgements This work was partly supported by the French Space Agency (Centre National d’Etudes Spatiales, CNES) through contract no. 8361. The authors would like to thank CNES, the Italian Space Agency (ASI), and GeoEye Inc. and Google Crisis Response for providing the Pléiades, COSMO-SkyMed, and GeoEye imagery used for experiments.
Publisher Copyright:
© Springer International Publishing AG 2018.