Deep Multimodal Hashing with Orthogonal Units

Deep Multimodal Hashing with Orthogonal Units

Deep Multimodal Hashing with Orthogonal Units Daixin Wang1, Peng Cui1, Mingdong Ou1, Wenwu Zhu1 1Tsinghua National Laboratory for Information Science and Technology Department of Computer Science and Technology, Tsinghua University. Beijing, China Abstract Faced with such huge volumes of multimodal data, multi- modal hashing is a promising way to perform efficient multi- Hashing is an important method for performing ef- modal similarity search. The fundamental problem of multi- ficient similarity search. With the explosive growth modal hashing is to capture the correlation of multiple modal- of multimodal data, how to learn hashing-based ities to learn compact binary hash codes. Despite the suc- compact representations for multimodal data be- cess of these hashing methods [Kumar and Udupa, 2011; comes highly non-trivial. Compared with shallow- Zhang et al., 2011; Zhen and Yeung, 2012], most existing structured models, deep models present superiori- multimodal hashing adopts shallow-structured models. As ty in capturing multimodal correlations due to their [Srivastava and Salakhutdinov, 2012a] argued, correlation of high nonlinearity. However, in order to make the multiple modalities exists in the high-level space and the learned representation more accurate and compact, mapping functions from the raw feature space to the high lev- how to reduce the redundant information lying in el space are highly nonlinear. Thus it is difficult for shallow the multimodal representations and incorporate d- models to learn such a high-level correlation [Bengio, 2009]. ifferent complexities of different modalities in the Recently, multimodal deep learning [Ngiam et al., 2011b; deep models is still an open problem. In this pa- Srivastava and Salakhutdinov, 2012a; 2012b] has been pro- per, we propose a novel deep multimodal hashing posed to capture the high-level correlation in multimodal in- method, namely Deep Multimodal Hashing with formation. [Kang et al., 2012] also proposed a multimodal Orthogonal Regularization (DMHOR), which ful- deep learning scheme to perform hashing. However, existing ly exploits intra-modality and inter-modality corre- works do not consider the redundancy problem in the learned lations. In particular, to reduce redundant infor- representation, which makes it incompact or imprecise, and mation, we impose orthogonal regularizer on the significantly affects the performance of efficient similarity weighting matrices of the model, and theoretical- search. In shallow-structured models, it is common to direct- ly prove that the learned representation is guaran- ly impose the orthogonal regularization on the global dataset teed to be approximately orthogonal. Moreover, we to decorrelate different bits. However, due to the high nonlin- find that a better representation can be attained with earity, the objective function of deep learning is highly non- different numbers of layers for different modali- convex of parameters, thus potentially causes many distinct ties, due to their different complexities. Compre- local minima in the parameter space [Erhan et al., 2010]. In hensive experiments on WIKI and NUS-WIDE, order to alleviate this problem, mini-batch training is com- demonstrate a substantial gain of DMHOR com- monly adopted for deep learning [Ngiam et al., 2011a], but pared with state-of-the-art methods. this intrinsically prohibits the possibility to impose regular- ization directly on the final output representation of global dataset. Therefore, how to solve the redundancy problem 1 Introduction of hashing representation learning by deep models is still a Nowadays, people have generated huge volumes of multi- challenging and unsolved problem. Furthermore, most of modal contents on the Internet, such as texts, videos and im- the previously proposed deep models adopt symmetric struc- ages. Multimodal data carries different aspects of informa- tures, with the assumption that different modalities possess tion, and many real-world applications need to integrate mul- the same complexity. However, it is intuitive that visual data timodal information to obtain comprehensive results. For ex- has much larger semantic gap than textual data, which result- ample, recommendation systems aim to find preferred multi- s in different complexities in different modalities. How to modal items (e.g. web posts with texts and images) for user- address the imbalanced complexity problem in deep learning s, and image search systems aim to search images for tex- models is also critical for multimodal hashing. t queries. Among these important applications, multimodal To address the above problems, we propose a Deep Mul- search, which integrates multimodal information for similar- timodal Hashing model with Orthogonal Regularization (as ity search is a fundamental problem. shown in Figure 1) for mapping multimodal data into a common hamming space. The model fully captures intra- Zhang and Li, 2014]. However, these multimodal hashing modality and inter-modality correlations to extract useful in- models adopt shallow-layer structures. It is difficult for them formation from multimodal data. In particular, to address to capture the correlation between different modalities. the redundancy problem, we impose orthogonal regularizer- Only a few recent works focus on multimodal deep learn- s on the weighting matrices, and theoretically prove that the ing. [Ngiam et al., 2011b; Srivastava and Salakhutdinov, learned representation is approximately guaranteed to be or- 2012a; 2012b] target at learning high-dimensional latent fea- thogonal. For the problem of imbalanced complexities, we tures to perform discriminative classification task. [Wang et empirically tune the number of layers for each modality, and al., 2014; Feng et al., 2014] apply autoencoder to perform find that a better representation can be attained with different cross-modality retrieval. However, these methods are all dif- number of layers for different modalities. ferent from our task of learning compact hash codes for multi- The contributions of our paper are listed as follows: modal data. From the angle of targeted problem, the most re- • We propose a novel deep learning framework to gener- lated work with ours is [Kang et al., 2012], which proposed a ate compact and precise hash codes for multimodal data, multimodal deep learning scheme to perform hashing. How- by fully exploiting the intra-modality and inter-modality ever, in their scheme, it does not consider the redundant in- correlation and incorporating different complexities of formation between different bits of hash codes. The redun- different modalities. dancy in hash codes will badly influence the performance of similarity search due to the compact characteristic of hashing • We propose a novel method with theoretical basis to representations. In addition, they fail to consider the different reduce the redundancy in the learned hashing repre- complexity of different modalities. sentation by imposing orthogonal regularization on the weighting parameters. 3 The Methodology • Experiments on two real-world datasets demonstrate a In this section, we present Deep Multimodal Hashing with substantial gain of our DMHOR model compared to Orthogonal Regularization (DMHOR) in detail and analyze other state-of-the-art hashing methods. its complexity to prove the scalability. 2 Related work 3.1 Notations and Problem Statement In recent years hashing methods have experienced great suc- In this paper, we use image and text as the input of two differ- cess in many real-world applications because of their supe- ent modalities without loss of generality. Note that it is easy riority in searching efficiency and storage requirements. In for the model to be extended to incorporate other forms of general, there are mainly two different ways for hashing to representations and more modalities. The terms and notations generate hash codes: data-independent and data-dependent of our model are listed in Table 1. Note that the subscript v ways. represents image modality and t represents text modality. Data-independent hashing methods often generate random projections as hash functions. Locality Sensitive Hashing Table 1: Terms and Notations (LSH) [Datar et al., 2004] is one of the most well-known Symbol Definition mv ,mt number of hidden layers in image or text pathway representative. It uses a set of random locality sensitive n number of samples hashing functions to map examples to hash codes. Further M the length of the hash codes [ ] xv , xt image or text input improvements such as multi-probe LSH Lv et al., 2007 (l) (l) hv , ht the l-th hidden layer for image or text pathway are proposed but the performance is still limited by the ran- h top joint layer dom projection technique. Data-dependent hashing methods (l) (l) Wv , Wt l-th layer’s weight matrix for image or text pathway (l) (l) were then proposed. They use machine learning to utilize bv , bt l-th biases for image or text pathway (l) (l) (l) mv +1 (l) (l) (l) mt+1 the distribution of data to help improve the retrieval qual- θ fWv ; bv ; cv gl=1 [ fWt ; bt ; ct gl=1 (l) (l) ity. Spectral Hashing [Weiss et al., 2008] is a representa- sv ,st l-th layer’s unit numbers for image or text pathway tive. Then some other hashing methods are proposed, in- cluding shallow structured methods [Norouzi and Blei, 2011; Suppose that we have n training examples with image- Liu et al., 2012; Wang et al., 2010] and deep learning based text pairs, represented by Xv = fxv;1; :::; xv;ng and Xt = methods [Salakhutdinov and Hinton, 2009; Xia et al., 2014]. fxt;1; :::; xt;ng. The main objective is to find M-bit binary Most of the above methods are designed for single modali- hashing codes H for training examples, as well as a corre- ty data. However, we often need to process multimodal in- sponded hash function f(·), which satisfies H = f(Xv;Xt). formation in real-world applications. Therefore, some re- Moreover, if two objects O1 and O2 are semantic similar, the cent works have focused on encoding examples represent- hash functions should satisfy that the distance of hash codes ed by multimodal features. For example, Cross-View Hash- H(O1) and H(O2) is small.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    7 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us