Learning-Shared Cross-Modality Representation Using Multispectral-LiDAR and Hyperspectral Data

Hleð...
Thumbnail Image

Dagsetning

Höfundar


Journal Title

Journal ISSN

Volume Title

Útgefandi

Institute of Electrical and Electronics Engineers (IEEE)

Úrdráttur

Due to the ever-growing diversity of the data source, multimodality feature learning has attracted more and more attention. However, most of these methods are designed by jointly learning feature representation from multimodalities that exist in both training and test sets, yet they are less investigated in the absence of certain modality in the test phase. To this end, in this letter, we propose to learn a shared feature space across multimodalities in the training process. By this way, the out-of-sample from any of multimodalities can be directly projected onto the learned space for a more effective cross-modality representation. More significantly, the shared space is regarded as a latent subspace in our proposed method, which connects the original multimodal samples with label information to further improve the feature discrimination. Experiments are conducted on the multispectral-Light Detection and Ranging (LIDAR) and hyperspectral data set provided by the 2018 IEEE GRSS Data Fusion Contest to demonstrate the effectiveness and superiority of the proposed method in comparison with several popular baselines.

Lýsing

Publisher's version (útgefin grein)

Efnisorð

Cross-modality, Feature learning, Hyperspectral, Multimodality, Multispectral-Light Detection and Ranging, Shared subspace learning, Fjarkönnun

Citation

Hong, D., Chanussot, J., Yokoya, N., Kang, J., Zhu, X.X., 2020. Learning-Shared Cross-Modality Representation Using Multispectral-LiDAR and Hyperspectral Data. IEEE Geoscience and Remote Sensing Letters. doi:10.1109/lgrs.2019.2944599

Undirflokkur