Introduction
Owing to different data collection methods and data types, multi-source heterogeneous data for the same thing can be described in a variety of different ways, modes, or perspectives. For the description of the same thing, some data are described with a coarse-grained structure, such as labels, while some data are described with a fine-grained structure, such as a scene graph. Therefore, the effective utilization of these massive multi-source heterogeneous data [1] containing rich semantic information, mining the deep internal connections of multi-source heterogeneous data, and providing users with high-quality cross-modal information retrieval services have become key issues in the current field of information retrieval.
Multimodal data contain rich multi-granularity semantic information, including coarse-grained [2] semantic conceptual features and fine-grained semantic features [3]. Mining multi-granularity semantic features can effectively learn the semantic correlations between multimodal data [4], thereby narrowing the “semantic gap” [5], which indicates the gap between low-level data features and actual semantics or concepts. For example, a gap exists between the pixel representation of an image and the conceptual or semantic information contained in the image. Resolving the semantic gap requires associating low-level features with higher-level semantics to better understand and retrieve cross-modal data.
Many existing cross-modal related studies have achieved good results on certain publicly available experimental datasets, but they still face some challenges and problems in practical applications, mainly because the representation information of images and text is not fully mined [6]. Traditional cross-modal retrieval models only consider the mining of global semantic information, without considering the objects and their relationships, resulting in some fine-grained semantic information being ignored. If global semantic information and fine-grained semantic information are fully considered and integrated, richer representation information can be obtained. Insufficient cross-modal representation learning [7]. Different modalities of data, such as images and text, may contain different types of information. Preserving inter-modal heterogeneity can ensure a wider range of information coverage and enrich the content and breadth of search results. Additionally, heterogeneity allows for complementarity between modalities, meaning that one modality may contain information that another modality does not possess. For example, in image retrieval, the image itself may convey certain visual features, whereas the relevant text may provide richer semantic information. This complementarity can improve the accuracy and comprehensiveness of the retrieval. Traditional cross-modal retrieval models are simply mapped to a common representation space and then subjected to cross-modal representation learning. This can lead to changes in the heterogeneity between modalities and invariance within modalities, resulting in a loss of correlation and a reduction in the performance of cross-modal retrieval. This requires finding effective representation learning methods, capable of integrating semantic information from different modal data and maintaining data similarity in a shared space. It is difficult to construct a good hash function [8], and many methods in reality cannot consider the nonlinear structure of the data, and instead use continuous solution methods in real space. This method cannot capture nonlinear projections and causes significant quantization losses, particularly when the hash code length is long.
In response to these shortcomings, we propose a scene graph model to perform hierarchical semantic association modeling on multimodal data. For each modality of the data, intra-modal global semantic mapping was used to represent coarse-grained semantic association information, whereas a relational semantic graph based on the scene graph was used to describe fine-grained semantic association information. This can provide sufficient semantic discrimination information for cross-modal quantization coding learning. When cross-modal retrieval is performed between images and text, the visual features of image data and semantic information of text data are heterogeneous, and their expression forms and features are different. Then, by enforcing comprehensive similarity preservation, we narrow the “heterogeneous gap” [9] between different domains and modalities. In addition, a direct learning approach for generating hash codes is introduced, which reduces the impact of quantization loss. This approach not only simplifies the learning process but also enhances the quality of the resultant hash codes. The main contributions of this study are as follows:
Multi-granularity semantic feature modeling. We propose a scene graph based multi-granularity semantic information preservation model. This model fully mines more fine-grained and coarse-grained semantic information within images and texts by integrating global semantic information and scene graph semantic relations, which can effectively narrow the low-level features with higher-level semantic gaps and provide a good foundation for hash encoding learning.
Comprehensive similarity preservation. We propose a triplet loss to reduce the heterogeneity between modalities in cross-modal representation learning and design a reconstruction loss to better capture feature representation and similarity measures within media types.
Direct Hash Learning. We propose a direct learning method for hash codes that reduces quantization loss and information loss caused by the inability to capture nonlinear semantic information.
Relate Work
A. Cross-Modal Retrieval
The motivation for cross-modal retrieval stems from the inherent heterogeneity of multimedia data and the limitations of traditional unimodal retrieval methods [10]. For example, text-based retrieval methods may not be able to capture the rich visual content in images or videos. By integrating multiple modalities, cross-modal retrieval helps to represent information more comprehensively and accurately, thereby improving the overall retrieval performance [11]. Figure 1 shows a basic schematic diagram of cross-modal retrieval.
The supervised cross-modal hash method is one of the most common types of cross-modal retrieval technique. In the past decade, many important approaches have been proposed and have achieved good retrieval performance. Kumar et al. [12] constructed a set of hash functions for multimodal data and transformed the hash function learning problem into a solvable eigenvalue problem using a new relaxation method. Liu et al. [13] proposed a flexible collaborative matrix decomposition hash algorithm (FS-CMFH) for efficient cross-modal retrieval that utilizes label consistency between different modalities to preserve semantic information within and between modalities in the common latent semantic representation space. Wang et al. [14] proposed a label consistency matrix decomposition hash method that maps heterogeneous data to a latent semantic representation space, in which multimodal data from the same category share the same feature representation. Jiang et al. [15] initially introduced a cross-modal hash model grounded in deep neural networks, amalgamating multimodal feature and hash learning within a single model. Yang et al. [16] integrated different types of paired constraints using a deep model to enhance the similarity measurement of hash codes from both intra-modal and inter-modal perspectives. Deng et al. [17] formulated a deep hash (TDH) using triplets as supervised information to foster semantic connections among cross-modal instances. The supervised hash method based on the deep learning model not only has strong nonlinear semantic representation ability, but also achieves end-to-end semantic feature learning and hash code generation, substantially enhancing cross-modal hash retrieval accuracy.
Although supervised cross-modal hashing methods have achieved a good retrieval performance, they rely heavily on annotated data. However, in massive multi-source heterogeneous data, the amount of annotated data is very limited, and annotating a large amount of data consumes huge manpower and resources, which makes the existing supervised hash methods difficult to apply to large-scale multi-source heterogeneous data [18]. Therefore, many researchers have focused on the research and exploration of unsupervised cross-modal hash methods.
To fully explore the potential correlation among multimodal data, Liu et al. [19] partitioned the feature space into a shared subspace for all hash functions and their complementary subspaces. This shared space captures a common structure across diverse hash functions, encapsulating crucial information for effective hash learning. Hu et al. [20] introduced an iterative multi-view hash algorithm (IMVH) aimed at acquiring optimal alignment within coding schemes to preserve similarity across views. Irie et al. [21] proposed an unsupervised hash method called alternating common quantization, which alternately searches for binary quantizers for each modal space by connecting multimodal data, ensuring the minimum quantization difference while maintaining data similarity. Su et al. [18] introduced an unsupervised deep joint semantic reconstruction hash (DJSRH) method. It amalgamates the distinct modality’s original neighborhood details using a semantic similarity matrix to encapsulate the inherent semantic relationships. Huang et al. [22] proposed an unsupervised cross-modal hash framework that utilizes data fusion to capture the underlying manifolds across modalities, avoiding the problem of maximizing mutual constraints between intra-modal and inter modal similarities. Yang et al. [23] investigated a deep semantically aligned hash algorithm (DSAH) that intelligently aligns feature similarity with hash code similarity via a semantic alignment loss function.
To effectively reduce the heterogeneity between modalities, some studies have integrated deep adversarial learning methods into cross-modal hash learning models, thereby improving the quality of hash code generation. For example, He et al. [24] devised an unsupervised cross-modal retrieval technique that leverage adversarial learning. The inclusion of a modal classifier to anticipate the modality of the feature transformation guarantees statistical indistinguishability among these transformed features. Zhang et al. [25] maximized the unsupervised representation learning potential inherent in generative adversarial networks (GANs) to explore the underlying structure of cross-modal data. They formulated an unsupervised cross-modal hash model based on GANs. At present, unsupervised methods lack semantic labels on the data, making it difficult for unsupervised hash models to effectively mine key semantic features and achieve cross-modal semantic alignment. Therefore, fully learning and characterizing the complex correlations between multi-source heterogeneous data and improving the semantic discriminability of multimodal unified hash codes have become key issues for the widespread application of unsupervised methods.
B. Scene Graph
A scene graph is a data structure in the fields of computer vision and graphics and is used to represent the relationships between objects in an image or scene. It represents objects in an image as nodes and uses edges to represent the relationships between objects. The main purpose of scene graphs is to capture the hierarchical and semantic relationships between objects in a structured manner to better understand and process the image content. In computer vision, this structured representation is helpful in tasks such as image understanding, object detection, and scene generation. The scene graph is composed of a triple form of “object-relationship-object” where “object” is a node and “relationship” is an edge connecting two objects. Yu et al. [26] proposed a method called DGCPN. The core idea of this method is to use the concept of graphs to establish associations between different modal data. The network structure proposed in this study aims to maintain the neighbor relationship of data in the graph, thereby maintaining data consistency when encoding data. This graph neighborhood consistency helps ensure that similar data remain similar in low-dimensional embedding spaces. Johnson et al. [27] explored a scene graph-based image retrieval method to improve the performance and accuracy of image retrieval by utilizing the scene graph information of images. Li et al. [28] introduced a method for the semantic modeling and representation of objects and their relationships in images by generating scene maps from object, phrase, and region descriptions. Figure 2 shows an image of the scene graph.
The image Scene Graph. This picture depicts the red team and the white team standing on a green field playing rugby, where “red team” “white team” “rugby” and “green field” are objects, and the relationship between them is indicated by two-way arrows; “standing” “playing” are relational words, and the relationship between them and the objects is indicated by single arrows.
C. Hash
In the era of big data, cross-modal hashing finds extensive applications in cross-modal retrieval. Its popularity stems from its high efficiency, low dimensionality, minimal storage requirements, and effective representation of high-level feature consistency in diverse multi-source data [29]. Current cross-modal hashing methods have not fully leveraged the abundant semantic details within vast and varied multi-source data. Consequently, the primary challenge in cross-modal hash retrieval research involves effectively exploring these multi-source heterogeneous data to inform the training of deep cross-modal hash models and enhance the quality of the generated hash code. A basic schematic of the cross-modal hash is shown in Figure 3.
In recent years, deep neural networks have substantially enhanced the performance of cross-modal hashing owing to their robust representational capabilities. Zhang et al. [7] introduced a cross-modal hash generation adversarial network model and utilized a reinforcement learning algorithm to guide its training. Shen et al. [30] proposed a semi-supervised graph convolutional hash network (SGCH) method that learns a common cross-modal Hamming space through end-to-end neural networks. Liu et al. [31] proposed a similarity hash (JDSH) method based on a joint modal distribution. This approach employs an unsupervised learning algorithm called Distributed Similarity Decision and Weighting (DSDW) to produce hash codes with increased discriminative capability. Ma et al. [32] introduced DASH in 2016, which is an uncomplicated yet impactful approach designed for cross-modal hashing. The DASH method, which is both non-iterative and devoid of parameters, is remarkably straightforward to implement and requires only three lines of code. They also published a paper called DCMH [33] in 2017, which assessed the efficacy of the proposed discrete optimization by optimizing its objective function using a relax-and-threshold technique. Extensive empirical evaluations conducted on image-text and image-tag datasets show that DCMH is a substantial advancement compared to prior methods, exhibiting notable enhancements in training efficiency and retrieval accuracy. The performance of semantic information retrieval is enhanced by the incorporation of joint supervision mechanisms, specifically involving instance-pairwise, instance-labeled, and class-wise similarities. Yu et al [34] further augmented the semantic consistency across heterogeneous modalities. The integration of these supervision strategies contributes to an improved and refined retrieval process that ensures a more comprehensive and accurate representation of semantic information in diverse data modalities. These achievements have theoretically verified the feasibility of cross-modal retrieval based on cross-modal hashing, but existing hash methods have not yet fully explored deep internal connections in multi-source heterogeneous data. Therefore, mining deep internal correlations in multi-source heterogeneous data from different granularities and modal data to improve the representation ability of hash codes on multi-source heterogeneous data is of great research value. This study aims to start with deep mining of different granularities data, different modal data, and the inherent correlation relationships between different data sources. Studied the representation of cross-modal hashes for multi-granularity semantic fusion to enhance the representation ability of cross-modal hashes for multi-source heterogeneous data.
Proposed Method
Our model consists of three parts: multi-granularity semantic feature modeling, comprehensive similarity preservice, and cross-modal hash learning. First, feature vectors of images and texts were obtained through convolutional neural networks [35] and bag-of- words [36] models. Then, a global semantic graph was constructed using these feature vectors, which saved coarse-grained semantic information regarding the image and text. Simultaneously, we constructed scene graph semantic information to integrate and connect objects and object relationships in images and texts, to retain more fine-grained semantic information. In this way, we obtained complete coarse-grained and fine-grained semantic information of images and texts, providing richer and more accurate semantic descriptions for cross-modal retrieval tasks. Second, when mapping image and text features to a common space [37], we used a triplet constraint [38] method to reduce the heterogeneity between modalities in cross-modal representation learning and designed the reconstruction loss to better capture feature representation and similarity measures within media types. Finally, in the learning process of hash functions, we propose a direct learning method for hash codes that reduces quantization loss and information loss caused by the inability to capture nonlinear semantic information. Figure 4 illustrates the framework of the proposed model.
A. Problem Formulation
Let D represent a multimedia dataset containing a set of images, denoted by the capital letter I, containing n images. We can then use sets \begin{equation*} \text {sim}(I, T) = \cos (\theta) = \frac {\phi (I) \cdot \psi (T)}{\|\phi (I)\| \cdot \|\psi (T)\|}, \tag{1}\end{equation*}
Here,
B. Global Semantics Relation Graph Construction
To realize coarse-grained semantic feature modeling, the Euclidean distance and kNN algorithms were proposed to measure the similarity between images and texts, respectively. where \begin{align*} E_{I:ij}&=\mathrm {EucDst}(\boldsymbol {C}_{i}, \boldsymbol {C}_{j}), \tag{2}\\ E_{T:ij}&=\mathrm {EucDst}(\boldsymbol {W}_{i}, \boldsymbol {W}_{j}), \tag{3}\end{align*}
C. Local Semantic Relation Graph Construction
To achieve fine-grained semantic feature modeling, the proposed approach constructs semantic entities and their semantic relationships in image data based on the local semantic relationship graphs of scene graphs. Specifically, we represent a visual scene graph as
To achieve fine-grained semantic feature modeling, a hierarchical local semantic graph based on a scene graph was proposed to construct semantic entities and their semantic relations in text data. Specifically, for the text data, the text is first segmented to obtain a word sequence
After the global semantic relation graph and local semantic relation graph are constructed, the semantic representations of the network nodes are first learned separately by a graph convolutional neural network, and then the node representations are fused by a semantic fusion network, which can obtain efficient semantic representations \begin{align*} \boldsymbol {\xi }_{i}&=\sigma \left ({\boldsymbol {M}_{f}^{I}\left ({\boldsymbol {g}_{i}^{I} \bowtie \boldsymbol {G}_{i}^{I}}\right)+\boldsymbol {\beta }^{I}}\right), \\ \boldsymbol {\zeta }_{t}&=\sigma \left ({\boldsymbol {M}_{f}^{T}\left ({\boldsymbol {g}_{i}^{T} \bowtie \boldsymbol {G}_{i}^{T}}\right)+\boldsymbol {\beta }^{T}}\right), \tag{4}\end{align*}
D. Modality Invariance Learning
We strive to maintain cross-modal invariance by minimizing the difference between representations of similar items across modalities, and simultaneously maximizing the distinction between dissimilar items within the same modality. To achieve this, we incorporate triplet constraints in embedding via a dedicated triplet loss term. The process of selecting positive and negative sets for triplet loss is as follows. Samples from different modalities but with the same label are composed of sample pairs, for example, let \begin{align*} L_{\text {i}} &= \sum _{i=1}^{N} \left [{ d(f(A_{i}), f(P_{t})) - d(f(A_{i}), f(N_{t})) + \alpha }\right]_{+} \tag{5}\\ L_{\text {t}} &= \sum _{t=1}^{N} \left [{ d(f(A_{t}), f(P_{i})) - d(f(A_{t}), f(N_{i})) + \alpha }\right]_{+} \tag{6}\\ L_{\text {triplet}} &= L_{\text {i}} + L_{\text {t}}, \tag{7}\end{align*}
E. Intra-Modal Similarity Learning
In the intra-modal similarity learning step, we focused on data similarity learning within each media type to better capture feature representations and similarity measurements within the media type. An autoencoder enables the input data to be encoded (compressed) and decoded (reconstructed) through a network. We designed a reconstruction loss to calculate the difference between the input of the autoencoder and the decoder output. By optimizing this loss, the autoencoder learns how to encode and decode the input data, thereby capturing their main features of the input data. The reconstruction loss can be expressed as \begin{equation*} L_{\text {reconstruction}} = \sum _{i=1}^{N} ||X_{i} - \hat {X_{i}}||_{2}^{2} \tag{8}\end{equation*}
Among them,
F. Hash Code Learning
After the above learning process, the semantic representations of each modality \begin{align*} \mathcal {L}_{HS} &= \gamma \left ({\| \boldsymbol {B}_{I} - [\boldsymbol {H}_{I}] \|_{F}^{2} + \| \boldsymbol {B}_{T} - [\boldsymbol {H}_{T}] \|_{F}^{2}}\right) \\ &\qquad \qquad \quad + \eta \left ({\| [\boldsymbol {H}_{I}] \mathbf {1} \|_{F}^{2} + \| [\boldsymbol {H}_{T}] \boldsymbol {1} \|_{F}^{2}}\right), \tag{9}\end{align*}
\begin{equation*} \mathcal {L} = \mathcal {L}_{HS} + L_{\text {reconstruction}} + L_{\text {triplet}}. \tag{10}\end{equation*}
Experiments
A. Dataset
In the experimental Section of this study, we provide a broad and in-depth evaluation of the performance of cross-modal retrieval. Experiments were performed on three distinct datasets: Wikipedia [39], NUS-WIDE [40], and MIRFlickr-25k [41]. These datasets cover image and text data from different domains and content types, providing a diverse set of challenges and evaluation environments. To improve the computational efficiency of our experiments, we selected only a part of the data in each dataset as the training and testing samples. A selection of examples from Wikipedia, NUS-WIDE, and MIRFlickr-25k is shown in Figure 5.
To assess the efficacy of the proposed method, we compared it with six representative cross-modal retrieval models. These models cover classical cross-modal retrieval methods. By comparing these models, we aimed to comprehensively evaluate the advantages of our approach under different datasets and model settings. We detail the experimental setup, performance metrics, and comparison results with each model in the following sections to demonstrate the effectiveness and superiority of our approach. The data for most of the comparison methods in the table below are from the original paper or are based on the source code provided in the paper. The following are concise views of these state-of-the-art technologies.
CVH [12] presented at ICJAI in 2011. Proposed a learned hash function method for cross-view similarity search. Maintaining similarity in the hash code space helps achieve efficient similarity matching in multi-view datasets, thus facilitating progress in multimodal data analysis.
CMFH [42] was presented at CVPR in 2014. Their main contribution is the attempt to use the Collective Matrix Decomposition (CMF) method to learn cross-view hash mapping relationships, which supports cross-view searching and enhances search accuracy by integrating information from multiple views.
DCMH [15] presented at the CVPR in 2017. The core idea of this study is to map data from different modalities to a shared low-dimensional binary code space, to achieve similarity matching of the cross-modal data in that space. This method utilizes the ability of deep neural networks to capture the correlations between modalities in high-dimensional feature spaces.
DJSRH [18] Presented at ICCV in 2019. This study proposes a joint reconstruction loss function that combines two key aspects: intra-modal reconstruction and inter-modal reconstruction. Effective retrieval of large-scale cross-modal datasets was achieved by utilizing the joint semantic information of the cross-modal data.
DGCPN [26] Presented at AAAI in 2021. The method utilizes graph embedding and deep learning to achieve effective hash encoding learning in cross-modal datasets by maintaining data neighborhood consistency. This helps to improve the performance and efficiency of multimodal data retrieval.
DFAH [43] Presented at Knowledge-Based Systems in 2022. A discrete fusion adversarial hashing method for cross-modal retrieval was proposed. It combines the ideas of discrete coding and adversarial training using adversarial training and fusion strategies, as well as the introduction of a discretized quantization loss.
B. Setting
To ensure a fair comparison, the proposed method is implemented using the DFAH framework. Similarity-preserving subnetwork layers were established with 4096 dimensions for both images and texts. Throughout the training, the parameters of the feature-extracting subnetworks remained fixed, with updates solely applied to similarity-preserving sub-networks. Employing a mini-batch SGD optimizer with a 0.9 momentum and weight decay of 0.0005, we utilized a mini-batch size of 32 with a learning rate set at 0.005. To determine the optimal hyperparameters
C. Performance Evaluation
To thoroughly assess the performance of our proposed cross-modal retrieval method, we used two main evaluation metrics: mean average precision (mAP) scores and precision-recall curves (PR-curves).
Mean average precision(mAP) is a commonly used performance metric that is particularly applicable to cross-modal retrieval tasks. In our experiments, mAP was used to measure the quality of retrieval results on different datasets. It considers the ranking of retrieved relevant texts/images and the relationships between them, thereby providing a more comprehensive performance evaluation. In calculating mAP, we focus on the average precision of each query (query) and then average the average precision of all queries to obtain the final mAP score. A high mAP score indicates that the retrieval results are of higher quality and ranked higher for relevant items.
Precision-recall curves are another method that is commonly used to evaluate retrieval tasks. In our experiments, we plotted PR curves to visualize the trade-off between different precision rates and recall rates. The precision rate represents the proportion of retrieved items that are truly relevant and the recall rate represents the proportion of retrieved truly relevant items out of all truly relevant items. By plotting the PR curves, we can see how the precision rate performs under different recall rates, which helps us understand the trade-off between the different performance metrics.
Using these two evaluation metrics, we comprehensively assessed the performance of our method across different datasets and model comparisons. The mAP scores provide an overall performance metric, while the PR-curves help us to gain insight into the performance at different operating points to better evaluate the superiority of our method.
The Wikipedia dataset is a relatively small cross-modal retrieval dataset with only 2866 image text pairs;however, the label classification inside is relatively abstract. Therefore, according to the settings in [44], we selected 2172 examples as training data, 462 examples as test data, and 231 examples as validation examples. From Table 2, we can see that the seven cross-modal retrieval methods, including ours, do not perform very well on the Wikipedia dataset compared with the other datasets. The reason for this lies in the greater abstractness of the categories within Wikipedia compared with the other two datasets, resulting in increased difficulty in acquiring discernible semantic features. However, our method still outperformed the other methods. Compared to the CVH and CMFH methods, our direct learning hash function strategy reflects advantages, indicating that our direct learning hash function strategy not only avoids the quantization loss but also maintains the similarity of similar objects in the binary space in a very good way. In addition, as shown Figure 7, the longer the hash code length, the better is the cross-modal retrieval performance, indicating that longer hash codes can reduce the impact of hash conflicts.
The NUS-WIDE dataset is an open-source cross-modal retrieval dataset that is very large, and some text descriptions are not only in English but also in other languages. Therefore, based on the ACMR settings [38], we selected 8000 samples as training data, 1000 samples as test data, and 1000 samples as validation examples. From Table 3, we can see that our method has better cross-modal retrieval results because we build a graph structure for tags and use graph strategies to explore the semantic dependencies. This is mainly because semantic knowledge is comprehensively and accurately represented by a multigraph so that more fine-grained semantic relationships can be mined to improve retrieval accuracy. This validates the importance of considering semantic connections among tags and highlights the value of investigating multiple tags as a promising avenue for research. Compared to the DCMH method, our method is more advantageous, because our intra-modal invariance strategy plays an important role in the common space mapping process. This strategy effectively reduced the heterogeneity between modalities in cross-modal representation learning, demonstrating the effectiveness of our method. However, based on Figure 8, our observation indicates that retrieving text from images yields superior results compared to retrieving images from text. This disparity stems from the fact that textual features capture the semantic essence of an object more effectively, whereas image features solely delineate the rudimentary visual attributes.
The MIRFLICKR-25K dataset contains many variations and a large amount of data. To ensure fairness in comparison, we selected 20015 samples as training data, 2000 samples as test data, and 2000 samples as validation examples based on the DFAH setting. As shown in Table 4, our method obtains better results than all other methods, compared to the CMFH method, which only uses coarse-grained features and ignores fine-grained complementary cues in cross-modal similarity learning. Our proposed method has a more powerful semantic representation capability, which means that it can capture more abstract conceptual information, and fully considered a combination of global coarse-grained semantic information and fine-grained local semantic information. In addition, from Figure 9, we can see that the DGCPN method also achieves relatively good results because the DGCPN method also considers fine-grained linguistic information, whereas the DGCPN method employs comprehensive similarity to prevent losses to pay attention to fine-grained linguistic information. DFAH, on the other hand, introduces the adversarial idea, which makes full use of semantic concepts to expand the interval distance between different semantic concepts, but does not consider inter-modal heterogeneity and intra-modal invariance.
D. Further Analysis
To further demonstrate the advantages of our proposed method in some aspects, we made some variations in the feature extraction preservation of image text and the learning of hash functions while keeping the other settings the same. MGSGH was the method proposed in this study. MGSGH-Nothing simply uses deep neural networks to extract features of images and texts, and then maps them to a binary common subspace for cross-modal retrieval. MGSGH-Only Graph fully explores the inherent correlation between data of different granularities by constructing a scene graph neural network, fusing coarse-grained and fine-grained representations, and then simply mapping them to a binary common subspace for cross-modal retrieval.
Table 5 lists the data with hash code lengths of 16, 32, and 64 bits obtained by applying the above three methods to the MIRFlickr-25k dataset. From this table, we can see that MGSGH achieves the best results, indicating that the construction of scene graph neural networks and direct learning of hash codes play important key roles. Scene graphs can provide rich correlation modeling for different object and relationship combinations, and help achieve more accurate and comprehensive cross-modal retrieval. According to the experimental data from the MGSGH and MGSGH-Only Graph, directly learning the hash code can effectively reduce quantization loss, retain more semantic information, and improve the efficiency of cross-modal hash retrieval.
Conclusion
This study introduces a scene graph model to perform hierarchical semantic association modeling on multimodal data. For each modality of data, coarse-grained semantic association information is represented using intra-modal global semantic maps, whereas fine-grained semantic association information is described using hierarchical local semantic maps based on scene graphs, providing sufficient semantic discrimination information for cross-modal quantization coding learning. By enforcing comprehensive similarity preservation, we establish effective cross-modal representations to bridge the differences in the feature space or data structure between image and text modal data. In addition, a direct learning approach for generating hash codes is introduced, which reduces the impact of quantization loss. This approach not only simplifies the learning process but also enhances the quality of the resultant hash codes. The proposed framework was rigorously validated on various benchmark datasets, thereby demonstrating its superior performance in retrieval tasks.