DOAJ Open Access 2025

Unsupervised Contrastive Hashing With Autoencoder Semantic Similarity for Cross-Modal Retrieval in Remote Sensing

Na Liu Guodong Wu Yonggui Huang Xi Chen Qingdu Li +1 lainnya

Abstrak

In large-scale multimodal remote sensing data archives, the application of cross-modal technology to achieve fast retrieval between different modalities has attracted great attention. In this article, we focus on cross-modal retrieval technology between remote sensing images and text. At present, there is still a large heterogeneity problem in the semantic information extracted from different modal data in the remote sensing field, which leads to the inability to effectively utilize intraclass similarities and interclass differences in the hash learning process, ultimately resulting in low cross-modal retrieval accuracy. In addition, supervised learning-based methods require a large number of labeled training samples, which limits the large-scale application of hash-based cross-modal retrieval technology in the remote sensing field. To address this problem, this article proposes a new unsupervised cross-autoencoder contrast hashing method for RS retrieval. This method constructs an end-to-end deep hashing model, which mainly includes a feature extraction module and a hash representation module. The feature extraction module is mainly responsible for extracting deep semantic information from different modal data and sends the different modal semantic information to the hash representation module through the intermediate layer to learn and generate binary hash codes. In the hashing module, we introduce a new multiobjective loss function to increase the expression of intramodal and intermodal semantic consistency through multiscale semantic similarity constraints and contrastive learning and add a cross-autoencoding module to reconstruct and compare hash features to reduce the loss of semantic information during the learning process. This article conducts a large number of experiments on the UC Merced Land dataset and the RSICD dataset. The experimental results of these two popular benchmark datasets show that the proposed CACH method outperforms the most advanced unsupervised cross-modal hashing methods in RS.

Penulis (6)

N

Na Liu

G

Guodong Wu

Y

Yonggui Huang

X

Xi Chen

Q

Qingdu Li

L

Lihong Wan

Format Sitasi

Liu, N., Wu, G., Huang, Y., Chen, X., Li, Q., Wan, L. (2025). Unsupervised Contrastive Hashing With Autoencoder Semantic Similarity for Cross-Modal Retrieval in Remote Sensing. https://doi.org/10.1109/JSTARS.2025.3538701

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber doi.org/10.1109/JSTARS.2025.3538701
Informasi Jurnal
Tahun Terbit
2025
Sumber Database
DOAJ
DOI
10.1109/JSTARS.2025.3538701
Akses
Open Access ✓