Next Article in Journal
Modeling of Duty-Cycled MAC Protocols for Heterogeneous WSN with Priorities
Next Article in Special Issue
A 3D Shape Recognition Method Using Hybrid Deep Learning Network CNN–SVM
Previous Article in Journal
Framework Integrating Lossy Compression and Perturbation for the Case of Smart Meter Privacy
Previous Article in Special Issue
Structure Fusion Based on Graph Convolutional Networks for Node Classification in Citation Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval

School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(3), 466; https://doi.org/10.3390/electronics9030466
Submission received: 14 February 2020 / Revised: 2 March 2020 / Accepted: 6 March 2020 / Published: 10 March 2020
(This article belongs to the Special Issue Deep Neural Networks and Their Applications)

Abstract

:
Multi-modal retrieval is a challenge due to heterogeneous gap and a complex semantic relationship between different modal data. Typical research map different modalities into a common subspace with a one-to-one correspondence or similarity/dissimilarity relationship of inter-modal data, in which the distances of heterogeneous data can be compared directly; thus, inter-modal retrieval can be achieved by the nearest neighboring search. However, most of them ignore intra-modal relations and complicated semantics between multi-modal data. In this paper, we propose a deep multi-modal metric learning method with multi-scale semantic correlation to deal with the retrieval tasks between image and text modalities. A deep model with two branches is designed to nonlinearly map raw heterogeneous data into comparable representations. In contrast to binary similarity, we formulate semantic relationship with multi-scale similarity to learn fine-grained multi-modal distances. Inter-modal and intra-modal correlations constructed on multi-scale semantic similarity are incorporated to train the deep model in an end-to-end way. Experiments validate the effectiveness of our proposed method on multi-modal retrieval tasks, and our method outperforms state-of-the-art methods on NUS-WIDE, MIR Flickr, and Wikipedia datasets.

1. Introduction

With the development of Internet and social media, people could come into contact with massive data from various modalities (e.g., image, text, audio, video, etc.) in daily lives. Multi-modal retrieval has become an important application scenario. For example, when a web user types into a textual description about “Obama”, the user usually expects to obtain related textual news of “Obama” and his images to best illustrate him. More image and text retrieval examples are shown in Figure 1. At the top in Figure 1, taking a text as query to retrieve the multi-modal dataset, related images, or texts are returned as the results, which are called text-to-image and text-to-text tasks respectively. At the bottom in Figure 1, taking an image as query, related texts or images are returned as the results, called image-to-text and image-to-image tasks, respectively. As we all know, traditional retrieval intra single modal data are often achieved by nearest neighboring search. However, heterogeneous data from different modalities are represented with distinct data distribution and different feature dimensions, thus they can’t be compared directly by traditional methods for such inter-modal retrieval tasks (text-to-image and image-to-text).
A straightforward solution is to model the joint distribution over the data of different modalities by projecting the heterogeneous data into a common subspace, in which heterogeneous data can be compared directly then multi-modal retrieval could be achieved. Typical methods learn two linear projections for two modalities respectively by maximizing the correlation [1,2] or minimizing the distance of one-to-one corresponding inter-modal data [3]. Nonlinear projections are further learned in deep learning models [4,5,6,7] with the corresponding relationship. Besides the correspondence, category information has been introduced into learning the discriminative common subspace to further narrow the gap between heterogeneous data in multi-view discriminant analysis [8,9,10] and canonical correlation analysis with clusters or labels [11,12,13]. These methods learn proper similarity or distance for multi-modal data. Most of them consider the pairwise correlation between corresponding data or instances belonging to the same class, but they don’t consider the correlation of dissimilar data directly.
Metric learning is to learn metric functions that satisfy pairwise must-link/cannot-link distance/similarity constraints or triplet-based distance/similarity constraints. It has been widely studied during the past few decades and has been developed in multi-modal retrieval. A lot of methods [14,15,16] have been proposed to learn common representations with linear metrics for multi-modal data by forcing the distances of inter-modal dissimilar data larger than similar data. Along with deep learning, deep metric learning has been developed in many visual understanding tasks [17], such as face recognition, image classification, visual search, visual tracking, person re-identification, and multi-modal matching. In the multi-modal task, hierarchical nonlinear transformations of deep neural network (DNN) [18,19,20,21] are utilized to learn the common multi-modal representations, in which the parameters of DNN for different modalities are optimized by modeling the similar and dissimilar constraints to preserve cross-modal relative ranking information. These methods are still based on hand-crafted image features, which are not optimal for multi-modal retrieval task since the stage of feature extracting and the stage of common representation learning are separated.
In recent studies [22,23,24], multi-modal data are embedded as low-dimensional representations via two different deep models and learned into a shared latent space at the top layers by pulling inter-modal similar points together while separating dissimilar points far away. They learn multi-layer nonlinear transformations to align heterogeneous data based on a binary similar or dissimilar inter-modal relationship, whereas they ignore the intra-modal correlation of multi-modal data, which has been verified effective in subspace learning methods [8,10,25,26,27]. In addition, a multi-scale semantic relationship is ignored in these methods, which consist of multi-scale similarities between inter-modal data and intra-modal data. Among inter-modal data, as shown in Figure 2, image 1 vs. text “blue sky and beach with rest area is a good place for vacation” are best matching, images 2 and 3 vs. the text are similar somewhat, whereas image 4 vs. the text are dissimilar. Among intra-modal data, as shown in Figure 2, images 1 vs. 2 are more similar than images 1 vs. 3; meanwhile, images 1, 2, and 3 are irrelevant to image 4.
In this paper, we propose Deep Multi-modal Metric Learning with Multi-Scale correlation (MS-DMML) for image and text retrieval. We design a deep model with two branches to learn common representations from raw heterogeneous data, where one branch is for image modality with stacked convolutional and fully-connected layers, and another branch is for text modality with stacked fully-connected layers. Inter-modal and intra-modal correlations are incorporated as loss function to train the deep model, in which the distances between both inter-modal and intra-modal instances are optimized to be consistent with multi-scale semantic relationship. In other words, image–text data with larger similarity are forced closer; meanwhile, dissimilar data are separated. Figure 3 illustrates the framework of our method. Experimental on three widely-used datasets are conducted and the results demonstrate the effectiveness of our MS-DMML method on four kinds of retrieval tasks, i.e., image-to-text, text-to-image, image-to-image, and text-to-text. The contributions are summarized as follows:
  • We propose a deep multi-modal metric learning method for image–text retrieval, in which two network branches are simultaneously learned as metric functions to measure the image–text distances according to multi-modal semantic relationship. Compared to most existing linear projection methods [2,3] and multilayer perception based on hand-crafted features [18,20], the proposed method effectively learns the comparable representations for heterogeneous data in an end-to-end way.
  • The semantic relationship between multi-modal data are formulated in a multi-scale similarity way as shown in Figure 2, rather than one-to-one correspondence [2,3] or binary similarity [15,22,23,24], which could force the deep network to learn fine-grained multi-modal distances and significantly improve retrieval performance.
  • We utilize the multi-scale semantic similarity to construct inter-modal and intra-modal correlations, which are incorporated to train the deep network. Our method benefits from combining inter-modal and intra-modal correlations, and achieves good performances on the four kinds of retrieval tasks.

2. Related Work

Our work is closely related to existing works on subspace learning, multi-modal metric learning, and deep learning for multi-modal retrieval. In this section, we briefly review the representative methods.
Typical methods in subspace learning is to learn two linear transformations for multi-modal data to project them into a common subspace, such as Canonical Correlation Analysis (CCA) [1], which projects heterogeneous data as comparable low-dimensional representations by maximizing the corresponding inter-modal correlation. CCA has many extensions and varieties [2,9,11,12,13]. Rasiwasia et al. [2] performed the classification on CCA learned representations and regarded the classification probabilities as the common features. Multi-view CCA [12] extends CCA to align the third view of high-level semantics with both image view and text view. Ranjan et al. [13] proposed multi-label CCA with multi-label annotations to form the semantic inter-modal relationship rather than one-to-one correspondence. Rasiwasia et al. [11] proposed cluster CCA to pair all points from one modality within a class with all points from the other modality in the same class and learn the projections using standard CCA. Sharma el al. [9] proposed generalized multi-view analysis with discriminative information by maximizing covariance between the cross-modal exemplars from the same class, which is a generalized and supervised extension of CCA. Partial Least Square (PLS) [28] learns the linear mapping from one modality to another modality such that the covariance between latent scores of the two modalities is maximized. Cross-modal Factor Analysis (CFA) [3] learns two orthogonal transformation spaces where pairwise inter-modal data are projected as close to each other as possible. These methods could be extended to kernel version to provide nonlinear feature representations, such as kernel CCA used for cross-modal retrieval by Hardoon et al. [29].
Besides the inter-modal correlation, there are several subspace learning methods utilizing intra-modal correlation when learning common representations. Song et al. [25] explored both inter-modal and intra-modal consistency to learn hash functions to map multi-modal data into the common Hamming space. Hu et al. [26] proposed iterative multi-view hashing to preserve the within-view similarity and the between-view correlations. Zhen et al. [27] proposed co-regularized hashing based on a boosted co-regularization framework with intra-modality and inter-modality loss. Kan et al. [8] formulated multi-view discriminant analysis by maximizing the between-class variations and minimizing the within-class variations of the learnt embeddings from both intra-view and inter-view perspectives. Jing et al. [10] proposed multi-view feature learning to process the supervised correlation information of samples with each view and between all views. Inspired by these shallow subspace learning methods, we explore both inter-modal and intra-modal correlations to learn deep distance metrics for heterogeneous data in this paper.
Numerous metric learning methods have been proposed in recent years, as a survey on metric learning for feature vectors and structured data in [30]. For example, Xing et al. [31] learnt a distance metric on points to respect their similar and dissimilar relationship. Large margin nearest neighbor classification method [32] learns a Mahalanobis distance metric with a goal that the k nearest neighbors always belong to the same class while examples from different classes are separated by a large margin. These methods are designed for a single kind of data; meanwhile, there has been a lot of metric learning methods for multi-modal data. A passive-aggressive model for image retrieval [14] method projects images into text space and ranks the relevant images appearing above the irrelevant images with text query by large margin learning. A bi-directional cross-media semantic representation model [15] optimizes the bi-directional list-wise ranking loss with a discriminative latent space embedding by the structural large margin learning. Huo et al. [16] proposed to optimize the performances under the receiver operating characteristic curve when learning the metric transformations for multi-modal data. Most of these methods learn linear distance metrics according to binary similarity constraints. However, we employ a deep learning method to learn multi-modal distances with multi-scale semantic constraints in this paper.
In recent years, deep learning has become an active research topic and shown strong nonlinear learning ability. Inspired by the improvement of deep neural network in single model retrieval, researchers have also attempted to apply deep models to multi-modal retrieval. Ngiam et al. [4] proposed a deep multimodal auto-encoder model to learn representations for coupled speech audio and videos of the lips. A deep Restricted Boltzmann Machine [5] is constructed by using separate latent models for each modality and a shared layer to learn the joint distribution of heterogeneous data. Andrew et al. [6] presented deep CCA, a deep learning method to learn complex projections for different modalities such that the resulting representations are highly correlated. Furthermore, Yan et al. [7] proposed an end-to-end learning scheme based on deep CCA. Feng et al. [33] utilized correspondence auto-encoder (Corr-AE) to correlate the hidden representations of two uni-modal auto-encoders. Wei et al. [34] proposed DeepSM to construct deep models for multi-modal data to learn the probabilities of classification as semantic common representations. These methods are typical deep multi-modal learning methods. However, they do not take into account dissimilar pairs directly, and thus tend to produce false positives in multi-modal retrieval tasks.
One recent direction emerging from deep learning community is adding the metric learning loss on top layers using the constraints on distances of heterogeneous data. Liong et al. [18] proposed deep coupled metric learning for cross-modal matching with multi-layer nonlinear transformations to map samples from different modalities into a shared latent feature subspace by minimizing the intra-class variation and maximizing the inter-class variation. Huang et al. [19] integrated quadruplet ranking loss and semi-supervised contrastive loss for modeling cross-modal semantic similarity in the deep model. Luo et al. [20] proposed learning to rank with nonlinear mapping functions for cross-modal data by using the self-paced learning with diversity. Peng et al. [21] adopted a two-stage learning based on deep neural network, in which intra-modal and inter-modal correlation are simultaneously modeled for feature learning and common representation learning stages, respectively. However, these methods are still based on hand-crafted features or features extracted with a trained deep model.
More recently, the common representations and feature extracting for multi-modal data are simultaneously learnt in an end-to-end deep metric learning model. Zhang et al. [22] constructed cross-modal ranking loss and within-modal discriminant loss to unify the training of two heterogeneous networks, deep residual network (ResNet) over images, and long short-term memory over sentences. Carvalho et al. [23] proposed a double-triplet scheme for both the retrieval loss and the class-based loss to train a dual deep neural network. Marin et al. [24] proposed to maximize the cosine similarity between positive recipe-image pair and minimize it between all non-matching recipe-image pairs, and incorporate additional semantic regularization through high-level classification. Wang et al. [35] proposed feature projector to generate common representations and confuse modality classifier, and they imposed triplet constraints on the feature projector. The losses for these methods are constructed based on binary similarity of inter-modal data (similarity and dissimilarity). Compared with these methods, multi-scale similarities are utilized to construct inter-modal and intra-modal semantic correlations, which are then combined to learn the deep distance metrics in this paper. The learned distances between multi-modal data would be fine-grained and reflect the multi-scale semantic relations conversely.

3. Deep Multi-Modal Metric Learning with Multi-Scale Correlation

In this section, we give the details of the proposed method. Two deep networks are designed as feature extracting and common representation learning for image and text modalities respectively, as shown in the left part in Figure 3, and distance constraints of inter-modality and intra-modality are constructed to be consistent with multi-scale semantic similarity, as shown in the right part in Figure 3.
The multi-modal dataset consists of two modalities with M image instances and N text instances, which is denoted as D = { D I , D T } . D I = { x m I , y m I | m [ 1 , M ] } denotes image data, where x m I represents the m-th raw image with three-channel RGB pixel values, and y m I is the category vector of the image and its dimension is the number of all categories in the dataset. D T = { x m T , y m T | m [ 1 , N ] } denotes text data, where x m T is bag-of-words representation and its dimension is the number of all words appearing in whole text dataset. y m T is category vector for the text. In contrast to CCA and its related methods [1,3,28], our solution does not need strictly one-to-one corresponding multi-modal data, thus M is not required to equal to N.

3.1. Deep Networks for Image and Text Modalities

Our network architecture consists of two branches that map each modality, i.e., image and text, into a common representation space, where the distances can be directly compared, as shown in Figure 3.
ResNet [36] promotes neural networks to a much deeper architecture and gains the accuracy from the considerably increased depth. Extensive works [22,23,24] have validated that ResNet can learn the most superior deep features for images in multi-modal retrieval tasks. Therefore, we apply ResNet-101 in an image branch to exploit the significant features implied in image data, which is pre-trained on ImageNet and fine-tuned in our experiments. On top of ResNet-101, we design stacked fully-connected layers to map image features into common representations, which will be compared with text representations. The outputs with 2048 dimensions before a classification layer of ResNet-101 are fed into the fully-connected network. All fully-connected layers are followed with nonlinear activation. The output of the last fully-connected layer is represented as f m I for the m-th image x m I .
In the text branch, the stacked fully-connected layers are utilized to learn feature representations. Input text is firstly represented as a bag-of-words vector and fed into the deep network. Compared to an image branch with a large number of convolution layers and stacked fully-connected layers, the deep architecture for text is simple. Convolution operation is usually used to extract image features due to it being possible to capture a strong spatial neighbor relationship. Text consists of many unordered words in some multi-modal retrieval scenarios, such as in datasets [37,38], thus we employ fully-connected operation rather than convolution to process bag-of-words vectors. Furthermore, semantic information is generally related with objects or categories, which are usually represented by nouns, that is, words could provide high-level semantic information relative to image pixels. Therefore, we believe that the relatively simple network is capable of extracting text representations for multi-modal retrieval. The fully-connected layers are followed with nonlinear activation, and the final output is denoted as f m T for text x m T .
The outputs of the last fully-connected layers for image and text are then normalized with
f L 2 , m I = f m I | | f m I | | ,
f L 2 , m T = f m T | | f m T | | ,
where the squared sums of f L 2 , m I and f L 2 , m T both equal to 1. The L 2 normalization operation is in favor of comparing the distances of multi-modal data on the same scale.

3.2. Multi-Scale Metric Learning

When learning deep metrics to project the heterogeneous data into the comparable representations, inter-modal and intra-modal correlations are built on semantic relations between multi-modal data. We propose to measure the semantic relations by the category vectors of multi-modal instances in a multi-scale similarity way, as shown in Figure 2. Inter-modal and intra-modal multi-scale similarity between the m-th instance and the n-th instance are defined as
S m , n = T r ( y m I ) y n T | | y m I | | | | y n T | | ,
S m , n I = T r ( y m I ) y n I | | y m I | | | | y n I | | ,
S m , n T = T r ( y m T ) y n T | | y m T | | | | y n T | | ,
where T r ( x ) denotes the transpose of vector x. S m , n is inter-modal semantic similarity, S m , n I is similarity for image modality, and S m , n T is for text similarity. In most existing work mentioned in Section 2, a semantic relationship between instances is binary (similar and dissimilar), where the similar instances share at least one category, while the dissimilar instances do not share any category. In our method, we refine the similar relationship to multi-scale and represent it with a real value (i.e., Equations (3)–(5)). If two instances are labelled with more common categories, the similarity value is closer to 1. Expect to learn the distance metrics between image and text modality so that the instances with larger semantic similarity value can be pulled closer to each other.
The loss of inter-modal multi-scale correlation is thus formulated as
m i n L I T = m = 1 M n = 1 N | | f L 2 , m I f L 2 , n T | | 2 S m , n s . t . | | f L 2 , p I f L 2 , q T | | 2 c , S p , q = 0 ,
where f L 2 , m I and f L 2 , n T are the normalized feedforward outputs of the deep networks of image and text modalities, respectively. If the inter-modal instances are with larger similarity S m , n , the square distance | | f L 2 , m I f L 2 , n T | | 2 should be smaller. It has a trivial solution with all metrics set to 0, thus a dissimilarity constraint is needed. If S p , q = 0 , which denotes that the image and the text are without common category labels, and the distance between the p-th image and the q-th text should be larger than a constant c. By minimizing L I T to optimize the deep model, the learned distances between inter-modal data are fine-grained and would be consistent with the multi-scale semantic constraint. However, not all dissimilar inter-modal instances could satisfy the hard constraints, so the slack variable is introduced into the inter-modal loss as
m i n L I T = m = 1 M n = 1 N α | | f L 2 , m I f L 2 , n T | | 2 S m , n + p = 1 M q = 1 N β ϵ p , q s . t . | | f L 2 , p I f L 2 , q T | | 2 c ϵ p , q , S p , q = 0 , ϵ p , q 0 ,
where α and β are the trade-off parameters between the multi-scale similarity and dissimilarity, and α + β = 1 . The inter-modal loss function is then equivalent to the following equation:
m i n L I T = m = 1 M n = 1 N α | | f L 2 , m I f L 2 , n T | | 2 S m , n + β ϵ m , n , ϵ m , n = m a x [ 0 , ( c | | f L 2 , m I f L 2 , n T | | 2 ) s i g ( S m , n ) ] ,
where function s i g ( x ) is set to 1 if x is equal to 0, otherwise s i g ( x ) is set to 0.
Besides the inter-modal correlation, intra-modal correlation among multi-modal data could promote the multi-modal retrieval performances, which is also formulated as metric learning with a multi-scale semantic relationship:
m i n L I = m = 1 M n = 1 M α | | f L 2 , m I f L 2 , n I | | 2 S m , n I + β ϵ m , n I , ϵ m , n I = m a x [ 0 , ( c | | f L 2 , m I f L 2 , n I | | 2 ) s i g ( S m , n I ) ] ,
where S m , n I represents the multi-scale similarity of the m-th image and the n-th image. By optimizing L I , more similar images are pulled closer and dissimilar images are separated. In addition, the same loss is defined for text modality as follows:
m i n L T = m = 1 N n = 1 N α | | f L 2 , m T f L 2 , n T | | 2 S m , n T + β ϵ m , n T , ϵ m , n T = m a x [ 0 , ( c | | f L 2 , m T f L 2 , n T | | 2 ) s i g ( S m , n T ) ] ,
where S m , n T represents the multi-scale similarity of the m-th text and the n-th text. In addition, more similar texts are pulled closer and dissimilar texts are separated by minimizing L T .
By combining Equations (8)–(10), we could get the following loss to train the two deep network branches for image and text modalities,
m i n L = λ 1 L I T + λ 2 L I + λ 3 L T ,
where λ 1 , λ 2 , and λ 3 are the trade-off parameters between the inter-modal and intra-modal correlation constraints, and λ 1 + λ 2 + λ 3 = 1 . By optimizing this weighted loss, our model could represent multi-modal data in a common space, in which their distances satisfy multi-scale semantic constraints. Thus, our method could reduce the heterogeneous and semantic gap between multi-modal data.
The optimization is to minimize the loss by fine-tuning ResNet and learning the multiple nonlinear transformations from sketch for image and text modalities. In training, we compute semantic similarity values S m , n , S m , n I and S m , n T by the given training dataset, and take the derivatives of objective function L with respect to f L 2 , m I and f L 2 , n T . The gradients are then back propagated to L 2 normalization, fully-connected layers and ResNet successively according to chain rule. We employ Adam [39] to update parameters with gradients. In tests, comparable representations are firstly extracted from raw images and texts with the trained deep model; then, the representations are used to achieve the four kinds of retrieval tasks by distance computing and ranking.

4. Experiments

In this section, we validate the effectiveness of our proposed method for inter-modal and intra-modal retrieval tasks on three public datasets NUS-WIDE [37] (https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE.html), MIR Flickr [38] (http://press.liacs.nl/mirflickr/), and Wikipedia [2] (http://www.svcl.ucsd.edu/projects/crossmodal/).

4.1. Dataset and Measurement

NUS-WIDE contains 269,498 images and corresponding tags, which are associated with 81 multi-label categories as ground truth, 1.8 on average for each image–text pair. Text annotations are represented with 1000-dimension bag-of-words vectors, and each text annotation has 7.8 words on average. We drop those pairs containing images without any category label or text annotation. We randomly select 20,000 image–text pairs for training and 2000 for validation from the training list provided by [37]. In addition, we randomly select 10,000 for test from the test list provided by [37].
MIR Flickr consists of 1 million images from the social photography website Flickr, where images are released under the Creative Commons License. Among the 1 million images, 25,000 have been annotated using 24 labels including categories such as bird, tree, people, indoor, sky, and night. A stricter labeling was done for 14 classes where an image was annotated with a category only if that category was salient. This leads to 38 multi-label categories, and 4.7 labels on average for each image. The images are all accompanied by user assigned tags on the Flickr website. We process all tag words and obtain the most frequent 2000 words as the bag to represent text modality, and each text has 6.3 words on average. We remove the images without tags or labels from the 25,000 images and 21,093 are remained. We random select 9093 for training, 2000 for validation, and 10,000 for test.
Wikipedia consists of 2866 image–text pairs from 10 categories such as history, art, biology, and literature. Each image and text belong to one category. We process all texts and obtain the most frequent 2000 words as the bag to represent text modality, and each text has about 106 words on average. We separate the training set 2173 pairs provided by [2] into 1973 for training and 200 for validation, and use 693 pairs provided by [2] for tests.
Mean Average Precision (mAP) is adopted for performance evaluation. The Average Precision (AP) of a query is computed by
A P = r = 1 R P ( r ) S ( r ) r = 1 R S ( r ) ,
where R is the number of retrieved results, P ( r ) denotes the precision of the top r retrieved items, S ( r ) = 1 if the r-th retrieved result is relevant, and S ( r ) = 0 , otherwise. mAP score is the mean of AP on all queries.

4.2. Implementation Details

For image modality, we utilize the pre-trained ResNet-101 to extract features from raw RGB images. The short edge of each image is firstly scaled to 256, and the long edge is proportionately scaled. The central region of images with size of 256 × 256 are then randomly cropped into 224 × 224 as the inputs fed into the deep model. The outputs with 2048 dimensions before classification layer are fed into fully-connected layers for learning the common representations with texts. For text modality, bag-of-words vectors are taken as inputs into fully-connected layers; thus, the input layer of text modality is with 1000 neural units for NUS-WIDE and 2000 for both MIR Flickr and Wikipedia. The numbers of fully-connected layers are both 2 for image and text branches. The number of neural units of hidden layer is 1024 and that of output layer is 256. Nonlinear activation function for the fully-connected layers is RELU. The outputs of images and texts are L 2 normalized and then used to compute the loss of inter-modality and intra-modality with multi-scale semantics. The parameters for all fully-connected layers are initialized with random values sampled from a normal distribution with mean of zero and standard deviation of 0.02 . Learning rate is set to 0.0001 . We train the model for 20 epochs with batch size 64. In a batch, we compute the correlations between any two samples; thus, 64 × 64 inter-modal and 2 × 64 × 63 intra-modal semantic correlation pairs are produced for training when processing the batch data to update the parameters of deep network. The parameter c in Equations (8)–(10) is set to 1. Other parameters λ 1 , λ 2 , λ 3 , α , and β are selected by grid search on validation datasets in the following experiments.

4.3. Validation on Inter-Modal and Intra-Modal Correlations

To validate the influence of inter-modal and intra-modal correlation on retrieval performances, we vary the value of parameter λ 1 from 0 to 1 with step 0.2 and set λ 2 = λ 3 in experiments. We report the representative mAP scores with R = 100 on validation datasets in Table 1, where α and β are fixed as 0.4 and 0.6 , respectively. The parameters with λ 1 = 0 , λ 2 = 0.5 , λ 3 = 0.5 mean that our model is individually trained with intra-modal correlation and λ 1 = 1 , λ 2 = 0 , λ 3 = 0 mean training with inter-modal correlation, thus called MS-DMML-Intra and MS-DMML-Inter, respectively. Our full model MS-DMML is set with the parameters λ 1 = 0.6 , λ 2 = 0.2 , λ 3 = 0.2 in this experiment.
On three datasets, the performances of MS-DMML-Intra on image-to-text and text-to-image are respectively 35.53 % , 19.90 % on NUS-WIDE, 64.12 % , 76.76 % on MIR Flickr, and 17.50 % , 14.60 % on Wikipedia. They are obviously lower than that of MS-DMML and MS-DMML-Inter, which indicates that the intra-modal correlation is not enough for inter-modal retrieval tasks. The performances of MS-DMML-Inter on image-to-text and text-to-image are respectively 57.41 % , 63.26 % on NUS-WIDE, 89.92 % , 87.08 % on MIR Flickr, and 46.55 % , 48.52 % on Wikipedia, which are as close as the performances of MS-DMML. It shows that the inter-modal correlation is useful for an inter-modal retrieval task. The performances of MS-DMML-Inter on text-to-text on the three datasets respectively 92.68 % , 80.32 % , 72.23 % , are lower than that of MS-DMML-Intra with 97.59 % , 82.79 % , 78.76 % , which shows that intra-modal correlation is suitable for intra-modal retrieval tasks. The performances of MS-DMML are, respectively 68.89 % , 89.27 % , 54.43 % on NUS-WIDE, MIR Flickr, and Wikipedia datasets on average, which are the best among MS-DMML-Inter and MS-DMML-Intra. This validates that the performances of multi-modal retrieval tasks benefit from combining the inter-modal and intra-modal correlations. All of the following experiments are based on λ 1 = 0.6 , λ 2 = 0.2 , λ 3 = 0.2 settings.

4.4. Validation Multi-Scale Correlation

In this section, we validate the effectiveness of multi-scale semantic correlation on validation datasets.
On NUS-WIDE, the value of multi-scale similarity (i.e., Equations (3)–(5)) for deep metric learning is between 0 and 1. The similarity of two instances without common category is represented as 0, and the similarity of two instances with one or more common categories is represented as larger than 0. However, the two cases are the binary semantic relationship in most existing methods [22,23,24]. To simulate the binary setting, we thus simplify our model by setting the similarity value as 0 if it is 0 and setting the value as 1 if it is larger than 0. The simplified model is called MS-DMML-bin. We compare the proposed full model with MS-DMML-bin and vary parameter α from 0.2 to 0.8 with step 0.2. The representative mAP scores with R = 100 are shown in Table 2. With α = 0.4 , β = 0.6 , the retrieval performances on four tasks of MS-DMML are respectively 59.51 % , 65.91 % , 52.55 % , 97.60 % , and the average value is 68.89 % , which is higher than the average performance of MS-DMML-bin with 67.15 % . With α = 0.8 , β = 0.2 , the average performance of MS-DMML ( 68.41 % ) is also higher than that of MS-DMML-bin ( 65.58 % ). The results show multi-scale semantic correlation can improve the performance of multi-model retrieval.
MIR Flickr is also a multi-label dataset, on which the semantic similarity setting is the same as that on NUS-WIDE, and we compare the full model MS-DMML with MS-DMML-bin on MIR Flickr. The mAP scores with R = 100 are shown in Table 2. With α = 0.4 , β = 0.6 , the retrieval performances on four tasks of MS-DMML are respectively 89.82 % , 88.68 % , 95.42 % , 83.18 % , and the average value is 89.27 % , which is higher than the average performance of MS-DMML-bin with 86.48 % . With α = 0.8 , β = 0.2 , the average performance of MS-DMML ( 83.80 % ) is also higher than that of MS-DMML-bin ( 80.52 % ). The results also show the effectiveness of multi-scale semantic correlation for promoting the multi-model retrieval.
Wikipedia is a single-label dataset; we thus compare it with the simplified version of our model, which only considers the correspondence of inter-modal data. If an image and text are one-to-one corresponding, the similarity is set to 1, otherwise 0. The simplified model is called MS-DMML-bin. The mAP scores with R = 100 are shown in Table 2. With α = 0.4 , β = 0.6 , the retrieval performances on four tasks of MS-DMML are respectively 47.44 % , 49.01 % , 42.94 % , 78.33 % , and the average value is 54.43 % , which is obviously higher than the average performance of MS-DMML-bin with 32.86 % . In addition, with α = 0.8 , β = 0.2 , the average performance of MS-DMML ( 54.56 % ) is also obviously higher than that of MS-DMML-bin ( 37.51 % ). The results show label information is really helpful for multi-modal retrieval, and our method could also process single-label data by incorporating the category information into similarity for learning semantic common representations.
From the results, we can obtain that the average performances of MS-DMML with α = 0.4 , β = 0.6 are close to α = 0.8 , β = 0.2 on NUS-WIDE and Wikipedia, and the average performance with α = 0.4 , β = 0.6 ( 89.27 % ) is higher than with α = 0.8 , β = 0.2 ( 83.80 % ) on MIR Flickr. This might be due to the number of average labels of MIR Flickr being more than that of NUS-WIDE and Wikipedia; therefore, the weight of similarity item α should be relatively small to form a trade-off between the multi-scale similarity and dissimilarity items. However, α = 0.4 , β = 0.6 for the three datasets could achieve a good performance; thus, we fix the parameters for all of the following comparison experiments.

4.5. Comparison with Others

MS-DMML is compared with several state-of-the-art methods, PLS [28], CCA [2], CFA [3], 3V-CCA [12], ml-CCA [13], DeepSM [34], and Recipe [24]. PLS, CCA, CFA, 3V-CCA, and ml-CCA are shallow correlation learning methods; thus, we use the 2048-dimensional visual features extracted with pre-trained ResNet-101 to learn the common representations for the five methods for fair comparison. The textual input of all methods is bag-of-words representations. DeepSM and Recipe methods are designed as end-to-end cross-modal learning frameworks, while the same deep neural networks are used in our experiments for fair comparison except following inevitable differences. DeepSM utilizes the classification probabilities of image and text data to form semantic mapping vectors for retrieval, where the dimensions of the learnt common semantic vectors are the number of categories, i.e., 81 for NUS-WIDE, 38 for MIR Flickr, and 10 for the Wikipedia dataset. The units numbers of the top fully-connected layers in deep neural networks are thus replaced with the corresponding 81, 38, and 10 for three datasets respectively when learning a DeepSM model. For all other methods, the common features learnt are of 256 dimensions. For PLS, CCA, CFA, 3V-CCA, ml-CCA, and DeepSM, inner product (IP) and Euclidean distance (ED) are both used as distance measurement to rank the retrieved data since their learnt representations are not normalized. For the Recipe method, cosine is used as distance measurement since cosine distances are used when learning the model in the paper [24]. For our method MS-DMML, inner product and Euclidean distance are equal since L 2 normalization is used before computing the inter-modal and intra-modal multi-scale constraints in our method. All experimental results are obtained on a test set of the three datasets.
The mAP performances with R = 100 of these methods on NUS-WIDE are reported in Table 3. The performances of three end-to-end deep methods, DeepSM, Recipe, and MS-DMML are better than that of PLS, CCA, CFA, 3V-CCA, and ml-CCA, which shows the designed fully-connected networks are effective to learn the common representations of multi-modal data in an end-to-end framework. 3V-CCA and ml-CCA are multi-label supervised methods, whose performances are better than PLS, CCA, and CFA, three correlation methods with one-to-one correspondence. Our method achieves 72.13 % mAP score on average of four retrieval tasks, which outperforms DeepSM’s performances, 64.75 % got with inner product and 65.81 % with Euclidean distance. This is due to 256-dimensional feature representations of our method stronger than 81-dimensional, classification representations obtained by DeepSM. Recipe uses classification loss as a regularization to learn the pairwise similarity/dissimilarity correlation between inter-modal data, whereas multi-scale similarity and intra-modal correlation are not considered in Recipe. This may be the reason that our method outperforms Recipe method with 72.13 % and 70.63 % average performances, respectively.
The mAP performances with R = 100 of these methods on MIR Flickr are reported in Table 4. The performances of three deep methods, DeepSM, Recipe, and MS-DMML are better than that of PLS, CCA, CFA, 3V-CCA, and ml-CCA, which also shows the effectiveness of end-to-end learning frameworks. 3V-CCA and ml-CCA are slightly lower than CCA and CFA. The anomalous performances might due to 3V-CCA and ml-CCA excessively focusing on the correspondence of multi-modal similar data (or the data of the same class), whereas ignoring the dissimilarity data, since category labels of MIR Flickr are relatively abundant to make an overdose of the similar data. Leveraging on metric learning, our method could adjust the proportion of dissimilarity constraint by parameter β , as the results of parameter experiment in Table 2, in which our model achieves better performance with β = 0.6 than with β = 0.2 . This validates the importance of dissimilar constraint, whereas it is ignored by most supervised subspace learning methods such as 3V-CCA and ml-CCA. Our method achieves an 89.49 % mAP score on average of four retrieval tasks, which outperforms DeepSM’s performances ( 82.59 % got with inner product and 84.83 % with Euclidean distance), and outperforms the Recipe method ( 86.79 % on average). This validates multi-scale semantic correlations of intra-modality and inter-modality are conducive to multi-modal retrieval.
The mAP performances with R = 100 of these methods on Wikipedia are reported in Table 5. The performances of deep methods, DeepSM, Recipe, and MS-DMML are better than that of PLS, CCA, CFA, 3V-CCA, and ml-CCA, which is the same as on NUS-WDIE and MIR Flickr. Supervised methods 3V-CCA and ml-CCA are better than PLS, CCA, and CFA on this single-label dataset. Our method achieves 58.97 % mAP score on average of four retrieval tasks, which outperforms DeepSM’s performances ( 53.31 % got with inner product and 49.99 % with Euclidean distance). This is due to 256-dimensional feature representations of our method stronger than 10-dimensional classification representations of DeepSM. Recipe achieves 55.34 % performance on average, which is worse than our method. This validates the semantic relationship and intra-modal correlation are helpful on single-label multi-modal dataset.

4.6. Performance Curves

We vary the number of retrieved results R from 50 to 500 with step 50 and plot the 10 mAP scores on the curves in Figure 4, Figure 5 and Figure 6 for three datasets. We compare MS-DMML with two end-to-end deep methods DeepSM and Recipe in this section. The red curves in the figures are mAP performances of our method MS-DMML. The black and green curves are mAP scores of DeepSM ranking with inner product and Euclidean distance. The blue curves are mAP scores of Recipe method.
In Figure 4, the performance curves on NUS-WIDE are plotted for four retrieval tasks. Sub-figure (a) is for an image-to-text retrieval task, from which we can see that the curve of MS-DMML is above others except at the first point with R = 50 . From sub-figures (b) and (c) for text-to-image and image-to-image retrieval tasks, we can obtain that our method outperforms DeepSM and Recipe. On sub-figure (d) for a text-to-text retrieval task, though the performances over 97 % are achieved by our method, the curve of MS-DMML is below that of Recipe and DeepSM with R > 200 . In Figure 5, the performance curves on MIR Flickr are plotted for four retrieval tasks. From the four sub-figures, we can see that all red curves are above others. It shows that MS-DMML outperforms DeepSM and Recipe methods on four kinds of multi-modal retrieval tasks. In Figure 6, the performance curves on Wikipedia are plotted for four retrieval tasks. From sub-figure (c) for an image-to-image task, we can see that the performance of MS-DMML is slightly higher than that of the Recipe method. From the other three sub-figures, we can see that our method outperforms DeepSM and Recipe methods prominently with R from 50 to 500. These curves validate that the performances of MS-DMML are robust regardless of how many retrieval results are considered for computing mAP.

4.7. Retrieval Examples

Examples of retrieval results on NUS-WIDE and MIR Flickr obtained with our method are shown in Figure 7 and Figure 8, respectively. The left side of each retrieval task is the query and the right side is the top 10 ranked results, and each textual result is listed at one line. Take the retrieval results on NUS-WIDE in Figure 7 as example. At the second line, querying with the image of semantic “food on table” are returned with 10 textual results, which almost contain relevant words “food, kitchen, orange, and cake” to the query. At the fourth line, querying with a textual document of “airplane military aircraft flying airport plane flight aviation fly landing”, the top 10 ranked images all well illustrate the textual information. At the fifth and sixth lines, querying with the images of semantic “food” and “cartoon yellow duck” respectively, the top 10 ranked images are all relevant; meanwhile, their appearances do not all look the same, which shows the effectiveness of modeling semantic correlation with supervised information in our method. Take the retrieval results on MIR Flickr in Figure 8 as example. At the second line, querying with a building image, the top 10 ranked results are listed on the right, and the fifth “fireworks” and sixth “hdr 500” are not relevant. At the fourth line, querying with “portrait light baby woman pregnant zen pregnancy mom”, the top 10 results are about “women” whereas not related to “pregnant”. This may be an improved direction to deal with irrelevant semantics in relevant data. At the fifth and sixth lines, querying with the images of semantic “flower” and “two persons walking” respectively, the top 10 ranked images are all relevant but not all the same at appearances, which is also observed on NUS-WIDE and shows that our model could retrieve diversified images with relevant semantics.

5. Conclusions

In this paper, we propose an end-to-end deep multi-modal learning framework for image–text retrieval. Two neural networks are designed to learn the comparable representations from raw data, and on top of them multi-scale semantic similarity and dissimilarity relations are constructed to form inter-modal and intra-modal correlation loss to learn the two metric networks. Multi-modal retrieval tasks are then achieved by nearest neighboring search. On NUS-WIDE, MIR Flickr, and Wikipedia datasets, our method outperforms state-of-the-art methods on image-to-text, text-to-image, image-to-image, and text-to-text retrieval tasks.

Author Contributions

Conceptualization, Y.H. and Y.Y.; methodology, Y.H.; software, Y.H.; validation, Y.H. and J.D.; formal analysis, Y.H.; investigation, Y.H.; resources, Y.H. and Y.Y.; data curation, Y.H.; writing—original draft preparation, Y.H.; writing—review and editing, Y.H. and Y.Y.; visualization, Y.H. and J.D.; supervision, Y.H.; project administration, Y.H.; funding acquisition, Y.H. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank the National Natural Science Foundation of China under Grant Nos. 61702466 and 61601414 for funding. This paper is also funded by the High-quality and Cutting-edge Disciplines Construction Project for Universities in Beijing (Internet Information, Communication University of China) (No. CUC190J051) and supported by the Fundamental Research Funds for the Central Universities.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hotelling, H. Relations between two sets of variates. In Breakthroughs in Statistics: Methodology and Distribution; Springer: New York, NY, USA, 1992; pp. 162–190. [Google Scholar]
  2. Rasiwasia, N.; Pereira, J.C.; Coviello, E.; Doyle, G.; Lanckriet, G.R.G.; Levy, R.; Vasconcelos, N. A new approach to cross-modal multimedia retrieval. In Proceedings of the 18th ACM International Conference on Multimedia (MM ’10), Firenze, Italy, 25–29 October 2010; pp. 251–260. [Google Scholar]
  3. Li, D.G.; Dimitrova, N.; Li, M.K.; Sethi, I.K. Multimedia content processing through cross-modal association. In Proceedings of the Eleventh ACM International Conference on Multimedia (MM ’03), Berkeley, CA, USA, 2–8 November 2013; pp. 604–611. [Google Scholar]
  4. Ngiam, J.; Khosla, A.; Kim, M.; Nam, J.; Lee, H.; Ng, A.Y. Multimodal deep learning. In Proceedings of the 28th International Conference on Machine Learning (ICML ’11), Bellevue, WA, USA, 28 June–2 July 2011; pp. 689–696. [Google Scholar]
  5. Srivastava, N.; Salakhutdinov, R. Multi-modal learning with deep boltzmann machines. In Proceedings of the Conference on Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 2222–2230. [Google Scholar]
  6. Andrew, G.; Arora, R.; Bilmes, J.; Livescu, K. Deep canonical correlation analysis. In Proceedings of the International Conference on Machine Learning (ICML’13), Atlanta, GA, USA, 16–21 June 2013; pp. 1247–1255. [Google Scholar]
  7. Yan, F.; Mikolajczyk, K. Deep correlation for matching images and text. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3441–3450. [Google Scholar]
  8. Kan, M.; Shan, S.; Zhang, H.; Lao, S.; Chen, X. Multi-view discriminant analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 1, 188–194. [Google Scholar] [CrossRef] [PubMed]
  9. Sharma, A.; Kumar, A.; Daume, H.; Jacobs, D.W. Generalized multiview analysis: A discriminative latent space. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2160–2167. [Google Scholar]
  10. Jing, X.Y.; Hu, R.M.; Zhu, Y.P.; Wu, S.; Liang, C.; Yang, J. Intra-view and inter-view supervised correlation analysis for multi-view feature learning. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence (AAAI ’14), Quebec City, QC, Canada, 27–31 July 2014; pp. 1882–1889. [Google Scholar]
  11. Rasiwasia, N.; Mahajan, D.; Mahadevan, V.; Aggarwal, G. Cluster canonical correlation analysis. In Proceedings of the Seventeenth International Conference on Artificial Intelligence and Statistics, Reykjavik, Iceland, 22–25 April 2014; pp. 823–831. [Google Scholar]
  12. Gong, Y.C.; Ke, Q.F.; Isard, M.; Lazebnik, S. A multi-view embedding space for modeling internet images, tags, and their semantics. Int. J. Comput. Vis. 2014, 2, 210–233. [Google Scholar] [CrossRef] [Green Version]
  13. Ranjan, V.; Rasiwasia, N.; Jawahar, C.V. Multi-label cross-modal retrieval. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4094–4102. [Google Scholar]
  14. Grangier, D.; Bengio, S. A discriminative kernel-based approach to rank images from text queries. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 8, 1371–1384. [Google Scholar] [CrossRef] [PubMed]
  15. Wu, F.; Lu, X.Y.; Zhang, Z.F.; Yan, S.; Rui, Y.; Zhuang, Y. Cross-media semantic representation via bi-directional learning to rank. In Proceedings of the 21st ACM International Conference on Multimedia (MM ’13), Barcelona, Spain, 21–25 October 2013; pp. 877–886. [Google Scholar]
  16. Huo, J.; Gao, Y.; Shi, Y.; Yin, H. Cross-modal metric learning for AUC optimization. IEEE Trans. Neural Netw. Learn. Syst. 2018, 10, 4844–4856. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Lu, J.; Hu, J.; Zhou, J. Deep metric learning for visual understanding: An overview of recent advances. IEEE Signal Process. Mag. 2017, 6, 76–84. [Google Scholar] [CrossRef]
  18. Liong, V.E.; Lu, J.; Tan, Y.; Zhou, J. Deep coupled metric learning for cross-modal matching. IEEE Trans. Multimedia 2017, 6, 1234–1244. [Google Scholar] [CrossRef]
  19. Huang, X.; Peng, Y. Cross-modal deep metric learning with multi-task regularization. In Proceedings of the 2017 IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 943–948. [Google Scholar]
  20. Luo, M.; Chang, X.; Li, Z.; Nie, L.; Hauptmann, A.G.; Zheng, Q. Simple to complex cross-modal learning to rank. Comput. Vis. Image Understand. 2017, 163, 67–77. [Google Scholar] [CrossRef] [Green Version]
  21. Peng, Y.; Qi, J.; Huang, X.; Yuan, Y. CCL: Cross-modal correlation learning with multi-grained fusion by hierarchical network. IEEE Trans. Multimed. 2018, 2, 405–420. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, L.; Ma, B.P.; Li, G.R.; Huang, Q.; Tian, Q. Multi-networks joint learning for large-scale cross-modal retrieval. In Proceedings of the 25th ACM international conference on Multimedia (MM ’17), Mountain View, CA, USA, 23–27 October 2017; pp. 907–915. [Google Scholar]
  23. Carvalho, M.; Cadene, R.; Picard, D.; Soulier, L.; THome, N.; Cord, M. Cross-modal retrieval in the cooking context: Learning semantic text-image embeddings. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (SIGIR ’18), Ann Arbor, MI, USA, 8–12 July 2018; pp. 35–44. [Google Scholar]
  24. Marin, J.; Biswas, A.; Ofli, F.; Weber, I.; Torralba, A. Recipe1M+: A dataset for learning cross-modal embeddings for cooking recipes and food images. IEEE Trans. Pattern Analy. Mach. Intell. 2019. [Google Scholar] [CrossRef] [Green Version]
  25. Song, J.; Yang, Y.; Yang, Y.; Huang, Z.; Shen, H.T. Inter-media hashing for large-scale retrieval from heterogeneous data sources. In Proceedings of the International Conference on Management of Data (SIGMOD ’13), New York, NY, USA, 23–28 June 2013; pp. 785–796. [Google Scholar]
  26. Hu, Y.; Jin, Z.; Ren, H.; Cai, D.; He, X. Iterative multi-view hashing for cross media indexing. In Proceedings of the 22nd ACM international conference on Multimedia (MM ’14), Orlando, Fl, USA, 3–7 November 2014; pp. 527–536. [Google Scholar]
  27. Zhen, Y.; Yeung, D.Y. Co-regularized hashing for multimodal data. In Proceedings of the 25th International Conference on Neural Information Processing Systems (NIPS ’12), Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1376–1384. [Google Scholar]
  28. Sharma, A.; Jacobs, D.W. Bypassing synthesis: PLS for face recognition with pose, low-resolution and sketch. In Proceedings of the CVPR 2011, Providence, RI, USA, 20–25 June 2011; pp. 593–600. [Google Scholar]
  29. Hardoon, D.R.; Shawe-Taylor, J. KCCA for different level precision in content-based image retrieval. In Proceedings of the Third International Workshop on Content-Based Multimedia Indexing, IRISA, Rennes, France, 22–24 September 2003. [Google Scholar]
  30. Bellet, A.; Habrard, A.; Sebban, M. A survey on metric learning for feature vectors and structured data. arXiv 2013, arXiv:1306.6709. [Google Scholar]
  31. Xing, E.P.; Ng, A.Y.; Jordan, M.I.; Rusell, S. Distance Metric Learning with Application to Clustering with Side-Information. In Proceedings of the 15th International Conference on Neural Information Processing Systems (NIPS ’02), Vancouver, BC, Canada, 9–14 December 2002; MIT Press: Cambridge, MA, USA, 2002; pp. 521–528. [Google Scholar]
  32. Weinberger, K.Q.; Saul, L.K. Distance metric learning for large margin nearest neighbor classification. J. Mach. Learn. Res. 2009, 10, 207–244. [Google Scholar]
  33. Feng, F.; Wang, X.; Li, R. Cross-modal retrieval with correspondence autoencoder. In Proceedings of the 22nd ACM international conference on Multimedia (MM ’14), Orlando, FL, USA, 3–7 November 2014; pp. 7–16. [Google Scholar]
  34. Wei, Y.C.; Zhao, Y.; Lu, C.Y.; Wei, S.; Liu, L.; Zhu, Z.; Yan, S. Cross-modal retrieval with CNN visual features: A new baseline. IEEE Trans. Cybern. 2017, 2, 449–460. [Google Scholar] [CrossRef] [PubMed]
  35. Wang, B.; Yang, Y.; Xu, X.; Hanjalic, A.; Shen, H.T. Adversarial cross-modal retrieval. In Proceedings of the 2017 ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 154–162. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Chua, T.S.; Tang, J.; Hong, R.; Li, H. NUS-WIDE: A real-world web image database from National University of Singapore. In Proceedings of the ACM International Conference on Image and Video Retrieval (CIVR ’09), Santorini Island, Greece, 8–10 July 2009; pp. 1–9. [Google Scholar]
  38. Huiskes, M.J.; Lew, M.S. The MIR Flickr retrieval evaluation. In Proceedings of the 1st ACM International Conference on Multimedia Information Retrieval (MIR ’08), Vancouver, BC, Canada, 30–31 October 2008; pp. 39–43. [Google Scholar]
  39. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. Illustration of multi-modal retrieval. At the top, the text is taken as a query to retrieve the multi-modal dataset, then ranking lists of semantic related images and texts are shown at right. At the bottom, the image is taken as a query to retrieve the multi-modal dataset, then ranking lists of semantic related texts and images are shown at right. The four kinds of retrieving tasks are called text-to-image, text-to-text, image-to-text, and image-to-image retrieval.
Figure 1. Illustration of multi-modal retrieval. At the top, the text is taken as a query to retrieve the multi-modal dataset, then ranking lists of semantic related images and texts are shown at right. At the bottom, the image is taken as a query to retrieve the multi-modal dataset, then ranking lists of semantic related texts and images are shown at right. The four kinds of retrieving tasks are called text-to-image, text-to-text, image-to-text, and image-to-image retrieval.
Electronics 09 00466 g001
Figure 2. Four images for multi-scale semantic correlation illustrations. In inter-modal data, the images 1, 2, and 3 vs. “blue sky and beach with rest area is a good place for vacation” from text modality are similar, but 1 vs. the text is the best matching, 2 takes the second place, and 3 is worse than both of them, whereas image 4 is dissimilar with the text. In intra-modal data, images 1 vs. 2 are more similar than images 1 vs. 3; meanwhile, images 1, 2, and 3 are irrelevant to image 4.
Figure 2. Four images for multi-scale semantic correlation illustrations. In inter-modal data, the images 1, 2, and 3 vs. “blue sky and beach with rest area is a good place for vacation” from text modality are similar, but 1 vs. the text is the best matching, 2 takes the second place, and 3 is worse than both of them, whereas image 4 is dissimilar with the text. In intra-modal data, images 1 vs. 2 are more similar than images 1 vs. 3; meanwhile, images 1, 2, and 3 are irrelevant to image 4.
Electronics 09 00466 g002
Figure 3. The framework of our proposed method MS-DMML. We design two deep neural networks for multi-modal data respectively, and raw image and text words are fed into the networks. The outputs are utilized to compute the distances of multi-modal data to respond to multi-scale correlation, and the whole networks are learned as metric functions for multi-modal retrieval.
Figure 3. The framework of our proposed method MS-DMML. We design two deep neural networks for multi-modal data respectively, and raw image and text words are fed into the networks. The outputs are utilized to compute the distances of multi-modal data to respond to multi-scale correlation, and the whole networks are learned as metric functions for multi-modal retrieval.
Electronics 09 00466 g003
Figure 4. Performance curves for the NUS-WIDE dataset. The horizontal axis is the number of retrieval results R, and the vertical axis is the mAP score. The four sub-figures are for four kinds of multi-modal retrieval tasks.
Figure 4. Performance curves for the NUS-WIDE dataset. The horizontal axis is the number of retrieval results R, and the vertical axis is the mAP score. The four sub-figures are for four kinds of multi-modal retrieval tasks.
Electronics 09 00466 g004
Figure 5. Performance curves for the MIR Flickr dataset.
Figure 5. Performance curves for the MIR Flickr dataset.
Electronics 09 00466 g005
Figure 6. Performance curves for the Wikipedia dataset.
Figure 6. Performance curves for the Wikipedia dataset.
Electronics 09 00466 g006
Figure 7. Retrieval examples from NUS-WIDE. For each retrieval example, query and its top 10 ranked results are listed.
Figure 7. Retrieval examples from NUS-WIDE. For each retrieval example, query and its top 10 ranked results are listed.
Electronics 09 00466 g007
Figure 8. Retrieval examples from MIR Flickr. For each retrieval example, query and its top 10 ranked results are listed.
Figure 8. Retrieval examples from MIR Flickr. For each retrieval example, query and its top 10 ranked results are listed.
Electronics 09 00466 g008
Table 1. mAP % on intra-modal and inter-modal correlation. (Bold numbers indicate the best performance).
Table 1. mAP % on intra-modal and inter-modal correlation. (Bold numbers indicate the best performance).
DatasetTasksMS-DMML-IntraMS-DMML-InterMS-DMML
NUS-WIDEImage-to-text35.5357.4159.51
Text-to-image19.9063.2665.91
Image-to-image51.6551.1152.55
Text-to-text97.5992.6897.60
Average51.1666.1168.89
MIR FlickrImage-to-text64.1289.9289.82
Text-to-image76.7687.0888.68
Image-to-image95.3293.5795.42
Text-to-text82.7980.3283.18
Average79.7587.7289.27
WikipediaImage-to-text17.5046.5547.44
Text-to-image14.6048.5249.01
Image-to-image42.8343.0142.94
Text-to-text78.7672.2378.33
Average38.4352.5854.43
Table 2. mAP % on multi-scale correlation. (Bold numbers indicate the best performance).
Table 2. mAP % on multi-scale correlation. (Bold numbers indicate the best performance).
DatasetTasksParameters α = 0.4 , β = 0.6 Parameters α = 0.8 , β = 0.2
MS-DMML-binMS-DMMLMS-DMML-binMS-DMML
NUS-WIDEImage-to-text56.1759.5155.5258.04
Text-to-image62.3765.9158.2863.36
Image-to-image53.2152.5554.6456.20
Text-to-text96.8697.6093.8996.04
Average67.1568.8965.5868.41
MIR FlickrImage-to-text85.5189.8279.3581.43
Text-to-image86.7988.6878.4184.11
Image-to-image93.5695.4287.5391.70
Text-to-text80.0683.1876.8077.97
Average86.4889.2780.5283.80
WikipediaImage-to-text28.5347.4433.8648.34
Text-to-image29.1849.0134.2549.41
Image-to-image34.8242.9438.0644.03
Text-to-text38.9178.3343.8876.45
Average32.8654.4337.5154.56
Table 3. mAP % on NUS-WIDE. (Bold numbers indicate the best performance).
Table 3. mAP % on NUS-WIDE. (Bold numbers indicate the best performance).
MethodsImage-to-TextText-to-ImageImage-to-ImageText-to-TextAverage
PLS(IP) [28]36.7537.2635.6448.2939.48
PLS(ED)21.8844.6051.3665.5845.85
CCA(IP) [2]53.0853.6751.5268.8056.77
CCA(ED)46.8647.6852.6572.9755.04
CFA(IP) [3]50.2356.2437.3284.5957.10
CFA(ED)53.4214.1650.6586.5851.20
3V-CCA(IP) [12]54.9857.8445.7182.7460.32
3V-CCA(ED)52.3221.1951.9084.8252.56
ml-CCA(IP) [13]46.7569.2255.0658.2957.33
ml-CCA(ED)50.9760.8053.0397.8965.67
DeepSM(IP) [34]46.1770.2244.0498.5664.75
DeepSM(ED)43.1167.3953.8798.8565.81
Recipe(cos) [24]60.7167.2655.1499.3970.63
MS-DMML61.3272.6355.4699.1372.13
Table 4. mAP % on MIR Flickr. (Bold numbers indicate the best performance).
Table 4. mAP % on MIR Flickr. (Bold numbers indicate the best performance).
MethodsImage-to-TextText-to-ImageImage-to-ImageText-to-TextAverage
PLS(IP) [28]60.6679.2584.9964.0372.23
PLS(ED)67.6776.8792.4272.6577.40
CCA(IP) [2]78.2478.0282.8778.3379.36
CCA(ED)73.6776.0781.9473.9776.41
CFA(IP) [3]84.6680.5391.4374.9582.89
CFA(ED)87.7965.1591.9072.0079.21
3V-CCA(IP) [12]76.2375.8090.5472.3978.74
3V-CCA(ED)81.6963.0491.2070.1576.48
ml-CCA(IP) [13]80.6479.8788.0777.9281.63
ml-CCA(ED)78.5671.5382.1375.6076.95
DeepSM(IP) [34]89.2574.6789.9976.4782.59
DeepSM(ED)87.5780.8892.7078.1984.83
Recipe(cos) [24]87.5183.0794.4282.1786.79
MS-DMML91.1787.6495.0284.1389.49
Table 5. mAP % on Wikipedia. (Bold numbers indicate the best performance).
Table 5. mAP % on Wikipedia. (Bold numbers indicate the best performance).
MethodsImage-to-TextText-to-ImageImage-to-ImageText-to-TextAverage
PLS(IP) [28]17.8425.5627.8121.5123.18
PLS(ED)27.6632.8842.0039.0935.41
CCA(IP) [2]28.3439.0840.0635.8935.84
CCA(ED)36.1428.6237.0547.7037.38
CFA(IP) [3]25.5738.6131.4351.4636.76
CFA(ED)35.5221.8840.2148.3436.49
3V-CCA(IP) [12]39.8843.5441.8157.9145.78
3V-CCA(ED)39.1530.2240.1448.0439.39
ml-CCA(IP) [13]36.7649.5546.1463.3750.20
ml-CCA(ED)46.9439.6342.3463.9148.20
DeepSM(IP) [34]45.4956.0539.8371.8553.31
DeepSM(ED)42.0747.4141.7068.7649.99
Recipe(cos) [24]47.7150.8349.6173.2255.34
MS-DMML49.0858.5350.6277.6458.97

Share and Cite

MDPI and ACS Style

Hua, Y.; Yang, Y.; Du, J. Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval. Electronics 2020, 9, 466. https://doi.org/10.3390/electronics9030466

AMA Style

Hua Y, Yang Y, Du J. Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval. Electronics. 2020; 9(3):466. https://doi.org/10.3390/electronics9030466

Chicago/Turabian Style

Hua, Yan, Yingyun Yang, and Jianhe Du. 2020. "Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval" Electronics 9, no. 3: 466. https://doi.org/10.3390/electronics9030466

APA Style

Hua, Y., Yang, Y., & Du, J. (2020). Deep Multi-Modal Metric Learning with Multi-Scale Correlation for Image-Text Retrieval. Electronics, 9(3), 466. https://doi.org/10.3390/electronics9030466

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop