Next Article in Journal
Route Configuration Method for Highway Passenger Hubs from the Perspective of Transportation Integration: A Case Study of Nanjing, China
Previous Article in Journal
Testing of a Virtualized Distributed Processing System for the Execution of Bio-Inspired Optimization Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Hashing with Gradient Attention

1
College of Software, Xinjiang University, Urumqi 830046, China
2
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(7), 1193; https://doi.org/10.3390/sym12071193
Submission received: 2 July 2020 / Revised: 12 July 2020 / Accepted: 14 July 2020 / Published: 18 July 2020

Abstract

:
The existing learning-based unsupervised hashing method usually uses a pre-trained network to extract features, and then uses the extracted feature vectors to construct a similarity matrix which guides the generation of hash codes through gradient descent. Existing research shows that the algorithm based on gradient descent will cause the hash codes of the paired images to be updated toward each other’s position during the training process. For unsupervised training, this situation will cause large fluctuations in the hash code during training and limit the learning efficiency of the hash code. In this paper, we propose a method named Deep Unsupervised Hashing with Gradient Attention (UHGA) to solve this problem. UHGA mainly includes the following contents: (1) use pre-trained network models to extract image features; (2) calculate the cosine distance of the corresponding features of the pair of images, and construct a similarity matrix through the cosine distance to guide the generation of hash codes; (3) a gradient attention mechanism is added during the training of the hash code to pay attention to the gradient. Experiments on two existing public datasets show that our proposed method can obtain more discriminating hash codes.

1. Introduction

Image retrieval is a hotspot in multimedia data retrieval, and there are many different methods today. The content-based image retrieval method [1] cannot meet the requirements of timeliness and retrieval precision due to its own limitations. In order to obtain higher retrieval accuracy, early researchers preferred the nearest neighbor search (NN). NN tends to retrieve the closest object from the dataset, and the complexity increases linearly with the amount of data, so its time performance cannot be guaranteed. Nowadays, there are a lot of high-dimensional multimedia data on social networks, the existing NN-based retrieval methods cannot obtain ideal retrieval results and an acceptable retrieval time. Therefore, researchers begin to pay attention to approximate nearest neighbor (ANN) [2,3,4,5,6,7,8,9]. Compared with the traditional retrieval method [1], another effective solution is the hashing method [2], through which high-dimensional media data can be converted into hash codes in the binary space.
In recent years, hashing has become a popular sub-region in ANN search [6,10,11]. With low storage cost and high query speed, the hash method has become the preferred method for effective ANN search. Hashing methods include two subcategories: (1) data-dependent hashing; (2) data-independent hashing. The process for data-independent hashing to obtain the hash code depends on the random mapping, and a reliable hash function cannot be obtained. Liu et al. [3] uses a set of hash functions to create multiple hash tables. After hash mapping, similar data points are more likely to collide in these hash tables, but the probability of collision of different points is very small. Therefore, similar images and non-similar images can be well distinguished. On this basis, Bai et al. [12] proposed to transfer the distance constraint of the hash code from the Hamming space to the Euclidean space. Data-dependent hashing methods use image colors, textures, shapes, or other features to directly generate hash codes and construct hash functions. Current researchers are committed to using deep learning methods to map features extracted from Convolutional Neural Network (CNN) to Hamming space through a hash function.
The data-dependent hashing method includes two subclasses: supervised hashing and unsupervised hashing. The difference between the two subclasses is whether the label information is used to construct the hash function. Gong et al. [13] was a representative algorithm in unsupervised hashing. The application of the Principal Components Analysis (PCA) algorithm could effectively reduce the dimensions of the original dataset, and at the same time constructed a binary hypercube. Each vertex of the hypercube represented the data point in the dataset, the quantization error was minimized by the rotation matrix. In supervised hashing methods [14,15,16,17,18,19], the construction process of the hash function in these methods used similar information of the paired images. In binary space, the Hamming distance of the image pair was determined by the image pair similarity.
The further development of CNN shows an excellent performance in feature extraction [20]. Existing supervised deep hashing methods have achieved a good performance [21,22,23,24,25,26]. However, for the massive image data existing in the network, it is not realistic to manually label them, and the hash method that relies on annotated data may not be suitable for practical applications.
In order to use these massive amounts of unlabeled data, the researchers proposed deep unsupervised hashing method. The learning process of the hash code in the early deep unsupervised hashing method was completed by the deep autoencoder [27,28]. However, the variability of natural images in terms of position, color, posture, etc., affected the representativeness of the learned hash code. Erin Liong et al. [29] maximized the representation of hash codes. Lin et al. [30] kept the hash code of the rotating image unchanged to learn hash codes. Jin et al. [31] further studied the rotation invariance between different images in the same category label. Yang et al. [32] used two half-Gaussian distributions to estimate the cosine distance distribution of different data points. Based on this hypothetical estimate, a pair-wise loss function was designed to retain semantic information. Generative adversarial networks (GAN) were widely used in various fields [33], Ghasedi Dizaji et al. [34] used a shared parameter generator and discriminator to optimize the hash function in the adversarial. On the basis of GAN, Deng et al. [35] constructed a semantic similarity matrix by using feature similarity and nearest neighbor similarity to guide the construction of hash codes.
The optimization algorithm commonly used in current deep learning models is gradient descent. Huang et al. [36] proposed that the hash model based on gradient descent optimization may fall into a dilemma during the optimization process, the hash codes of the paired images will always be updated in the direction of the other side, and the worst case is that the paired hash codes will exchange directions after being updated. This phenomenon is the “dilemma” of gradient descent. After falling into this dilemma, the Hamming distance corresponding to hash code will not be reduced, affecting the learning process, since this problem is caused by gradient changes in the process of updating hash codes in backpropagation. Selectively ignoring or reducing the gradient of one of the hash codes during backpropagation is an effective way to solve this problem. The gradient attention mechanism is designed to focus on the gradient of the paired image hash codes in supervised hashing.
Existing unsupervised deep hashing methods usually use a pre-trained network model to construct a similarity matrix to guide the generation of hash codes. Therefore, the above problems still exist in unsupervised learning, through the experimental results show that after adding gradient attention mechanism, unsupervised hash method achieved better performance.

2. UHGA Mmethod

2.1. Deep Hash Model

For unsupervised hashing, X = { x i } i = 1 N represents a dataset containing N images, where x i is the i-th image. Unsupervised hashing aims to obtain the binary code for each data point in the dataset and construct a hash function M ( θ ; x ) to generate this binary code, where θ represents the parameters of the deep model. Finally, after each image x i passes through the deep model M , it will be mapped into a k-dimensional vector Z = { z i } i = 1 n , where z i the i-th k-dimensional vector. After performing a b i = s i g n ( z i ) binarization operation on this k-dimensional vector, we obtain its corresponding hash code { b i { ± 1 } k } i = 1 n , where b i the i-th hash code. The sign function is expressed as:
s i g n ( x ) = { 1   i f   x 0 , 1   i f   x < 0
Unsupervised hashing usually guides the learning of hashing codes by constructing the semantic similarity matrix between image pairs. In this paper, we extract the features of FC-7 layer from n randomly selected data points in the dataset through the pre-trained Alexnet network to construct the similarity matrix. Most of the current deep hash algorithms use VGG-F network in MATLAB as the benchmark. In this paper, our experiment is implemented on the deep learning framework Pytorch. To ensure the fairness of the experiment, Alexnet, which has the same structure as VGG-F network, is selected as the benchmark network. For the input N data points, the corresponding feature matrix is obtained after the pre-trained Alexnet. According to the feature matrix, calculating the cosine distance of the corresponding feature of any two data points, and the cosine distance matrix D is constructed.
D = F i T F j F i × F j
where F i represents the feature of FC-7 layer.
Then, the similarity matrix S is constructed as follows through the cosine distance matrix:
S i j = { 1 , i f   d ( i , j ) < D . m e a n α , 0 , o t h e r s , 1 , i f   d ( i , j ) > D . m e a n + β
where D . m e a n represents the average value of cosine distance matrix D , α and β are threshold parameters, S i j = 1 means that x i is semantically similar to x j , S i j = 1 means the opposite, and S i j = 0 means that the semantic relationship between x i and x j is uncertain.
In the training process, the s i g n function cannot perform gradient training, t a n h is generally used instead to relax each binary code into a continuous hash code h [ 1 , + 1 ] k , where h k and b k denote the k-th bit of h and b .
To preserve the similarity of the paired image in the hash code, we use the inner product of the hash code to construct prediction similarity matrix :
i j = h i T h j
where h i and h j are relaxed hash codes.
To maintain the semantic information between image pairs, we minimize the L2 distance ( i j S i j ) 2 between the similarity matrix constructed by the feature matrix and the predicted similarity matrix. The resulting objective function is:
min θ l ( θ ) = 1 n 2 i = 1 n j = 1 n | S i j | ( i j S i j ) 2
where i j = 1 k h i T h j , h i = t a n h ( M ( θ ; x ) ) and θ is the parameter of the network model.
The commonly used method is to update the parameter θ by calculating the gradient of the loss function. First we define ν i = M ( θ ; x ) , the derivative of ν i to θ can be directly calculated, and then calculate the derivative of l ( θ ) to ν i by Equation(6).
l ( θ ) ν i = 2 k n 2 i = 1 n ( 1 h i 2 ) ( j = 1 n | S i j |   ( 1 k h i T h j S i j ) h j + ( 1 k h i T h j 1 ) h i )
Gradient descent is usually used to optimize parameters θ θ μ l θ . In the optimization process, the backpropagation method is used to calculate the gradient of the model parameter θ , the partial derivative l / is first calculated, and then the partial derivative h i k can be calculated by Equation (7).
l h i k = j l i j i j h i k
where h i k denote the k-th bit of h i .
Finally, the gradient l / θ can be calculated by (8):
l θ = i l h i h i θ
where l h i = [ l h i 1 , l h i 2 , , l h i k ] T .

2.2. The Dilemma of Gradient Decline

The purpose of network model optimization is to update the parameter θ of the model in the training process, but after updating θ , the detailed changes in h cannot be displayed directly, in order to analyze the change process of h , [36] proposes a lemma to analyze the update behavior of h .
Lemma 1:
for a given composite function g ( y ) , where x is a vector and y = f ( x ) is a scaling function.
Suppose x is updated by gradient descent, that is x + = x α g x , where α is the step size.
Conclusion: y + = f ( x + ) y a ^ g y , where a ^ is some positive scaler. i.e., y update along negative gradient.
Proof: 
Calculate the first-order Taylor expansion of f ( x + ) on x :
y + y = f ( x + ) f ( x ) y x , x + x = α y x , g x = α y x , g y , y x = α y x 2 g y = a ^ g y
where a ^ = α y x 2 . This proof is completed. □
By substituting x = θ , y = h and f ( x ) = t a n h ( M ( θ ) ) into Equation (9), it can be seen that when θ is updated by gradient descent, h changes along its negative gradient direction, and then the change in h can be observed by gradient.
For a pair of input images ( x i , x j ) , assume x i is similar to x j , i.e., s i j = 1 . If one of the bits in the current corresponding continuous hash code are h i = 1 and h j = 1 respectively, then the loss can be defined as = 1 i j = 1 h i T h j . At this time, > 0 , the current hash code needs to be updated. The gradients of h i and h j are:
l h i = l i j i j h i = h j
l h j = l i j i j h j = h i
Based on the lemma above, the updated values of h i and h j are expressed as h i + and h j + respectively, which can be expressed as:
h i + = h i α i j l h i = h i + α i j h j
h j + = h j α j i l h j = h j + α j i h i
where α i j and α j i are some positive scalers.
It can be seen from Equation (12) and (13) that, in the process of optimization, h i moves to h j direction and tends to have the same value as h j . The update of h j is the same as h i . Note that when h i = h j , i j = S i j , the hash codes are optimal. However, in some cases i.e., h i = h j = 1 ,   i j = S i j , and h i T h j = 1 when α i j = α j i = 2 , h i + = h j = 1 and h j + = h i = 1 , their inner product is still h i + T h j + = 1 after updating, therefore, the updated loss remains unchanged. In the next round of update, h i and h j will switch their symbols, and their corresponding losses will not be reduced, which will happen repeatedly. This dilemma also occurs when two different pairs of code are updated to stay away from each other.
Updating one of the paired hash codes during backpropagation is an effective solution, when their similarity prediction is not correct. When the prediction is correct, h i and h j are updated towards their sign value, which can effectively avoid the problem of symbol switching. When updating h i and h j the weight of the gradient should be carefully selected so as to update only one code when the similarity of prediction image is not correct, and two codes when the similarity of prediction is correct.

2.3. Gradient Attention Network

In this paper, we use gradient attention network to focus on the gradients of h i k and h j k , and generate weights for the derivatives of h i k and h j k before backpropagation to θ . We use the weighted derivatives of h i k and h j k instead of the original derivatives to calculate the gradients of θ . For all training pairs, the weighted derivatives of h i k and h j k are calculated as follows:
l h i k = j ξ i j , i k l i j i j h i k
l h j k = i ξ i j , j k l i j i j h j k
where ξ i j , i k and ξ i j , j k are the weights generated by the derivative of a pair of binary bits h i k and h j k respectively. A new gradient g ( φ ) can be obtained by substituting (14), (15) into (8). The parameters of the gradient attention network are expressed by φ .
For the hash model optimized by gradient descent, the change in the value of the corresponding hash code h i k of the paired image ( x i , x j ) is updated by the derivative related to the loss based on the original value during the optimization process. For the gradient attention network, all the inputs are h i k , l i j i j h i k and l h i k .
The gradient attention network first generates the feature value y i j , i k for the k-th bit of the hash code corresponding to the paired image, then calculates the weight of the k-th bit and normalizes it by the softmax function. The calculation process is as follows:
ξ i j , i k = e x p ( y i j , i k ) exp ( y i j , i k ) + e x p ( y i j , j k )
ξ i j , j k = exp ( y i j , j k ) exp ( y i j , i k ) + exp ( y i j , j k )
Figure 1 shows the deep unsupervised hash network model. The gradient attention network consists of two fully connection layers, each of which contains 100 hidden units.
The detailed learning process of the UHGA model is shown in Algorithm 1.
Algorithm 1. UHGA
Input:
 Training set X = { x i } i = 1 n ;
 Code length k;
Output:
 Updated network parameters θ and hash codes B .
Initialization:
 Initialize the Alexnet;
 Initialize parameters θ from the pre-trained Alexnet.
 Extracting features of Alexnet FC-7 layer.
 Using (2) to calculate cosine distance of paired image features.
 Using (3) to construct similarity matrix S .
Repeat
 a mini batch of images from X = { x i } i = 1 n , and for each image x i , follow each step below:
 •Calculate l ( θ ) and obtain the derivative l / in the process of backpropagation;
 •Using h i k , l i j i j h i k and l h i k as input of gradient attention network to calculate attention weights y i j , i k ;
 •Calculate ξ i j , i   k and ξ i j , j k by (16) and (17);
 •Calculate the new derivatives l h i k = j   ξ i j , i   k l i j i j h i k and obtain gradient of θ as g ( φ ) in the process of backpropagation;
 •Update θ by (18);
 •Calculate the loss l ( θ α g ( φ ) ) after update;
 •Update φ ;
 • θ θ + ;
Until number of iterations reached

2.4. Optimization

The gradient attention network can solve the problem of the exchanging positions of consecutive hash codes corresponding to pairs of images after updating. Due to this problem, the loss after updating remains unchanged, as the performance of attention weight can be evaluated by the loss reduced during the updating process. In gradient attention networks, the new gradient is first generated and then the parameter θ of the hash model is updated by Equation (18).
θ + = θ α g ( φ )
The updated loss is l ( θ + ) = l ( θ α g ( φ ) ) , and the reduced loss is l ( θ ) l ( θ + ) . After parameter θ is updated, the goal is still to minimize the loss; the goal for optimizing gradient attention is:
min φ l ( θ α g ( φ ) )
For Equation (19), θ can be optimized using Stochastic Gradient Descent (SGD) from Equation (6) and, as long as θ + is a function of g ( φ ) , φ can be optimized by any optimizer.

3. Experimental Results and Analysis

The UHGA model is based on the Alexnet implementation in Pytorch. Detailed configuration information for the experimental machine is shown in Table 1.

3.1. Datasets, Evaluation Metrics and Benchmarks

There are two authoritative image datasets: CIFAR-10 and NUS-WIDE datasets. The experiments have used these two datasets as benchmark datasets to compare with other methods in order to verify the feasibility of the UHGA model.
(1)
CIFAR-10. A total of 60,000 images, including 10 categories, and each category contains 6000 images, which is a single-label dataset.
(1)
NUS-WIDE. A total of 269,648 images, including 21 categories, each category is associated with at least 5000 images and each image belongs to one or more categories, which is a multi-label dataset.
In this article, the methods of SH (Spectral hashing) [37], SPH (Spherical hashing) [38], ITQ (Iterative quantization) [13], DSH (Density Sensitive Hashing) [39], SSDH (Semantic structure-based unsupervised deep hashing) [32], HashGAN (Unsupervised Deep Generative Adversarial Hashing Network) [34], and USDH (Unsupervised Semantic Deep Hashing) [31] are selected for comparison. We use Mean Average Precision (MAP), Precision–Recall curves (PR) and the precision curves of the first 1000 retrieval results (P@N) as evaluation metrics.

3.2. Hyperparameter Analysis

We use Equation (3) to construct a similarity matrix between images. There are two hyperparameters α and β in Equation (3). In order to analyze the influence of hyperparameters, we simply set the two hyperparameters to α = η ( D . m e a n D . m i n ) and β = η ( D . m a x D . m e a n ) . Then, we only need to adjust the hyperparameter η to observe the influence of the constructed similarity matrix on the retrieval accuracy. Table 2 shows the effect of different η on two datasets under different bits. In order to observe the change in MAP, we visualize it in Figure 2. Figure 2 shows that the value of MAP is relatively smooth when η is between 0.2-0.4, and when η continues to increase, MAP shows the opposite trend. Therefore, in subsequent experiments, the hyperparameter η is set to 0.3.

3.3. The effect of Gradient Attention

In order to study the effect of gradient attention, we added the gradient attention mechanism to the method based on pairwise loss. We used four methods in Table 3 as comparative experiments: SSDH, SSDH+GA, Ours, Ours+GA, where SSDH is its original method, SSDH+GA means adding gradient attention mechanism in SSDH, Ours means only using our own designed loss function to train the unsupervised hash model, and Ours+GA means adding a gradient attention mechanism to our own loss function.
It can be seen from Table (3) that, after adding the attention mechanism, the MAP of SSDH [32] is significantly improved. We obtained good results using our own designed loss function, and the map value continued to increase after using the attention mechanism. Figure 3 shows the decreasing trend of training loss before and after adding the gradient attention (GA) mechanism in SSDH [32] and our method. It can be clearly seen that the loss decreases faster after adding the gradient attention mechanism, which helps to obtain a better hash code.
The loss function we proposed in Table 3 has been greatly improved in MAP. This is also shown in the decreasing trend of training loss in Figure 3. The loss function we designed has more advantages in the generation of constrained hash codes. After adding the attention mechanism, our method and SSDH [31] are significantly improved on MAP. In Figure 3, the improvement in hash performance is reflected in the faster decline in training loss.
Figure 4 shows the effect of different learning rates on retrieval results. It can be seen from (a) and (b) in Figure 4 that, without the gradient attention mechanism, the retrieval quality will decline with the increase in the learning rate because, in the process of reverse propagation, when the learning rate is high, the network cannot converge. After using the gradient attention mechanism, the magnitude of the gradient descent is not only controlled by the learning rate, but also by the gradient attention. In the process of backpropagation, the gradient attention mechanism can select the appropriate gradient modification parameters for the network to obtain a better deep hash model.
The above situations indicate that the unsupervised method based on pairwise loss have a dilemma in gradient descent, and the design of the loss function also affects the probability of this dilemma. Obviously, the loss function we designed has a lower probability of dilemma. The addition of the gradient attention mechanism can effectively reduce the occurrence of this dilemma, thereby improving the performance of the hash method.

3.4. Comparison of Experimental Results

In the training of the hash function, we select 500 images from each image category to form a training dataset, and select 100 images for each category to serve as the test dataset on CIFAR-10. All remaining image samples are used as retrieved data. We select 21 kinds of common images from the NUS-WIDE dataset. Each class is associated with at least 5000 images, with a total of 190,000 images. The NUS-WIDE test dataset consists of 100 images of each class and 2100 images in total. The training dataset contains 5000 randomly selected images, and the rest of the images are retrieved as the sample dataset.
Table 4 shows the comparison results of all comparative experiments on the evaluation metric MAP. We calculated the MAP value of the hash code length of all the comparison methods from 16 bits to 128 bits on the two datasets. In the measurement results of different hash code length, our method is 8.73%, 8.67%, 12.36%, 13.79% higher than USDH in the CIFAR-10 dataset, and 5.77%, 5.71%, 9.51%, 9.21% higher than USDH in the NUS-WIDE dataset.
Figure 5and Figure 6 show the PR and P@N curves on CIFAR-10, respectively. From Figure 5, the area under the PR curve corresponding to our method is the largest, and our method shows better results as the length of the hash code increases. As can be seen in Figure 6, as the number of retrieval results continues to increase, the retrieval precision of all methods shows a downward trend, but our method always shows the best results.
Figure 7 and Figure 8 show the PR and P@N curve on NUS-WIDE respectively. Figure 7 shows that our method still has higher performance on multi-label datasets. Although the precision of our method is lower than USDH when the recall is between 0-0.1, the area under the PR curve corresponding to our method is still the largest overall; this also shows that our method is more stable. Figure 8 shows that the n curve is not as obvious as the downward trend in Figure 6, and our method still consistently shows the best performance in retrieval accuracy.
The above experimental results show that our unsupervised hash method achieves the best performance. Compared with other methods, the advantages of our proposed method are as follows: (1) the designed loss function can generate a hash code with a greater performance (2) by adding a gradient attention mechanism to reduce the impact of the dilemma in gradient descent during training.

4. Conclusions

In unsupervised deep hashing, a similarity matrix is usually constructed by extracting image features and used in the training of neural networks. Gradient descent is used to update parameters during training. However, for unsupervised hashing methods based on paired image training, there may be a problem in the hash code update process, as the hash codes corresponding to paired images are updated toward each other. After the network parameters are updated, the corresponding hash code changes its location, but the overall loss remains unchanged. This paper presents a Deep Unsupervised Hashing with Gradient Attention (UHGA) algorithm based on the deep learning end-to-end model. UHGA uses the cosine distance of features to construct similarity information between images, and the gradient attention mechanism in the process of gradient descent, which successfully obtains a better deep hash model and significantly optimizes the retrieval quality. Experiments on two datasets show that UHGA significantly improves the latest hash methods.

Author Contributions

Conceptualization, L.W. and S.J.; methodology, Y.L. and S.C.; software, S.J.; validation, A.D., Y.L. and L.W.; formal analysis, Y.L. and S.C.; writing—original draft preparation, S.J.; writing—review and editing, L.W.; visualization, A.D.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Xinjiang Uygur Autonomous Region Natural Science Foundation Project (2020D01C058), National Natural Science Foundation Project (61771416 and U1903213), Xinjiang Uygur Autonomous Region Higher Education Innovation Project (XJEDU2017T002), and Xinjiang Uygur Autonomous region graduate research and innovation project (XJ2019G039).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Y.; Zhang, D.; Lu, G.; Ma, W.Y. A survey of content-based image retrieval with high-level semantics. Pattern Recognit. 2007, 40, 262–282. [Google Scholar] [CrossRef]
  2. Tang, J.; Li, Z.; Wang, M.; Zhao, R. Neighborhood discriminant hashing for large-scale image retrieval. IEEE Trans. Image Process. 2015, 24, 2827–2840. [Google Scholar] [CrossRef] [PubMed]
  3. Paulevé, L.; Jégou, H.; Amsaleg, L. Locality sensitive hashing: A comparison of hash function types and querying mechanisms. Pattern Recognit. Lett. 2010, 31, 1348–1358. [Google Scholar] [CrossRef] [Green Version]
  4. Andoni, A.; Indyk, P. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions. In Proceedings of the Annual IEEE Symposium on Foundations of Computer Science (FOCS’06), Berkeley, CA, USA, 22–24 October 2006; Volume 47, pp. 459–468. [Google Scholar]
  5. Zhang, D.; Wang, J.; Cai, D.; Lu, J. Self-taught hashing for fast similarity search. In Proceedings of the 33rd international ACM SIGIR Conference on Research and Development in Information Retrieval, Geneva, Switzerland, 19–23 July 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 18–25. [Google Scholar]
  6. He, K.; Wen, F.; Sun, J. K-means hashing: An affinity-preserving quantization method for learning binary compact codes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OG, USA, 25–27 June 2013; pp. 2938–2945. [Google Scholar]
  7. Shen, F.; Zhou, X.; Yang, Y.; Song, J.; Shen, H.T.; Tao, D. A fast optimization method for general binary code learning. IEEE Trans. Image Process. 2016, 25, 5610–5621. [Google Scholar] [CrossRef] [PubMed]
  8. Xu, H.; Wang, J.; Li, Z.; Zeng, G.; Li, S.; Yu, N. Complementary hashing for approximate nearest neighbor search. In Proceedings of the 2011 International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 1631–1638. [Google Scholar]
  9. Wang, J.; Liu, W.; Kumar, S.; Chang, S.F. Learning to hash for indexing big data—A survey. Proc. IEEE 2015, 104, 34–57. [Google Scholar] [CrossRef]
  10. Wang, J.; Zhang, T.; Sebe, N.; Shen, H.T. A survey on learning to hash. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 769–790. [Google Scholar] [CrossRef]
  11. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef] [Green Version]
  12. Bai, X.; Yang, H.; Zhou, J.; Ren, P.; Cheng, J. Data-dependent hashing based on p-stable distribution. IEEE Trans. Image Process. 2014, 23, 5033–5046. [Google Scholar] [CrossRef] [Green Version]
  13. Gong, Y.; Lazebnik, S.; Gordo, A.; Perronnin, F. Iterative quantization: A procrustean approach to learning binary codes for large-scale image retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 2916–2929. [Google Scholar] [CrossRef] [Green Version]
  14. Shen, F.; Shen, C.; Liu, W.; Shen, H.T. Supervised Discrete Hashing. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 37–45. [Google Scholar]
  15. Cao, Z.; Long, M.; Wang, J.; Yu, P.S. Hashnet: Deep learning to hash by continuation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5608–5617. [Google Scholar]
  16. Yang, H.F.; Lin, K.; Chen, C.S. Supervised learning of semantics-preserving hash via deep convolutional neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 437–451. [Google Scholar] [CrossRef] [Green Version]
  17. Liu, H.; Wang, R.; Shan, S.; Chen, X. Deep supervised hashing for fast image retrieval. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2064–2072. [Google Scholar]
  18. Li, W.J.; Wang, S.; Kang, W.C. Feature learning based deep supervised hashing with pairwise labels. In Proceedings of the 2015 International Joint Conference on Artificial Intelligence (IJCAI), Buenos Aires, Argentina, 28 July–1 August 2015; pp. 1711–1717. [Google Scholar]
  19. Li, Q.; Sun, Z.; He, R.; Tan, T. Deep supervised discrete hashing. In Advances in Neural Information Processing Systems; Curran Associates: New York, USA, 2017; pp. 2482–2491. [Google Scholar]
  20. Wang, D.; Wang, L. On OCT image classification via deep learning. IEEE Photonics J. 2019, 11, 1–14. [Google Scholar] [CrossRef]
  21. Wu, Y.; Sun, Q.; Hou, Y.; Zhang, J.; Zhang, Q.; Wei, X. Deep covariance estimation hashing. IEEE Access 2019, 7, 113223–113234. [Google Scholar] [CrossRef]
  22. Cakir, F.; He, K.; Bargal, S.A.; Sclaroff, S. Hashing with mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2424–2437. [Google Scholar] [CrossRef] [Green Version]
  23. Li, N.; Li, C.; Deng, C.; Liu, X.; Gao, G. Deep joint semantic-embedding hashing. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 2397–2403. [Google Scholar]
  24. Cheng, S.; Wang, L.; Du, A. An adaptive and asymmetric residual hash for fast image retrieval. IEEE Access 2019, 7, 78942–78953. [Google Scholar] [CrossRef]
  25. Du, A.; Wang, L.; Cheng, S.; Ao, N. A Privacy-Protected Image Retrieval Scheme for Fast and Secure Image Search. Symmetry 2020, 12, 282. [Google Scholar] [CrossRef] [Green Version]
  26. Cheng, S.L.; Wang, L.J.; Huang, G.; Du, A.Y. A privacy-preserving image retrieval scheme based secure kNN, DNA coding and deep hashing. Multimed. Tools Appl. 2019. [Google Scholar] [CrossRef]
  27. Li, J. Context-based image re-ranking for content-based image retrieval. In Proceedings of the 2011 IEEE Symposium on Computational Intelligence for Multimedia, Signal and Vision Processing, Paris, France, 11–15 April 2011; pp. 39–46. [Google Scholar]
  28. Salakhutdinov, R.; Hinton, G. Semantic hashing. Int. J. Approx. Reason. 2009, 50, 969–978. [Google Scholar] [CrossRef] [Green Version]
  29. Erin Liong, V.; Lu, J.; Wang, G.; Moulin, P.; Zhou, J. Deep hashing for compact binary codes learning. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2475–2483. [Google Scholar]
  30. Lin, K.; Lu, J.; Chen, C.S.; Zhou, J. Learning compact binary descriptors with unsupervised deep neural networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1183–1192. [Google Scholar]
  31. Jin, S.; Yao, H.; Sun, X.; Zhou, S. Unsupervised semantic deep hashing. Neurocomputing 2019, 351, 19–25. [Google Scholar] [CrossRef] [Green Version]
  32. Yang, E.; Deng, C.; Liu, T.; Liu, W.; Tao, D. Semantic structure-based unsupervised deep hashing. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 1064–1070. [Google Scholar]
  33. Li, C.; Wang, L.; Cheng, S.; Ao, N. Generative Adversarial Network-Based Super-Resolution Considering Quantitative and Perceptual Quality. Symmetry 2020, 12, 449. [Google Scholar] [CrossRef] [Green Version]
  34. Ghasedi Dizaji, K.; Zheng, F.; Sadoughi, N.; Yang, Y.; Deng, C.; Huang, H. Unsupervised deep generative adversarial hashing network. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Coimbatore, India, 21–22 September 2018; pp. 3664–3673. [Google Scholar]
  35. Deng, C.; Yang, E.; Liu, T.; Li, J.; Liu, W.; Tao, D. Unsupervised semantic-preserving adversarial hashing for image search. IEEE Trans. Image Process. 2019, 28, 4032–4044. [Google Scholar] [CrossRef]
  36. Huang, L.K.; Chen, J.; Pan, S.J. Accelerate learning of deep hashing with gradient attention. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 5271–5280. [Google Scholar]
  37. Weiss, Y.; Torralba, A.; Fergus, R. Spectral hashing. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2009; pp. 1753–1760. [Google Scholar]
  38. Heo, J.P.; Lee, Y.; He, J.; Chang, S.F.; Yoon, S.E. Spherical hashing. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 2957–2964. [Google Scholar]
  39. Jin, Z.; Li, C.; Lin, Y.; Cai, D. Density sensitive hashing. IEEE Trans. Cybern. 2013, 44, 1362–1371. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Deep Unsupervised Hashing with Gradient Attention (UHGA) network structure.
Figure 1. Deep Unsupervised Hashing with Gradient Attention (UHGA) network structure.
Symmetry 12 01193 g001
Figure 2. MAP on different η. (a) MAP on CIFAR-10; (b) MAP on NUS-WIDE.
Figure 2. MAP on different η. (a) MAP on CIFAR-10; (b) MAP on NUS-WIDE.
Symmetry 12 01193 g002
Figure 3. (a) Training loss on CIFAR-10; (b) training loss on NUS-WIDE.
Figure 3. (a) Training loss on CIFAR-10; (b) training loss on NUS-WIDE.
Symmetry 12 01193 g003
Figure 4. MAP on different learning rates (lr): (a) MAP on CIFAR-10; (b) MAP on NUS-WIDE.
Figure 4. MAP on different learning rates (lr): (a) MAP on CIFAR-10; (b) MAP on NUS-WIDE.
Symmetry 12 01193 g004
Figure 5. PR on CIFAR-10. (a) PR with code length 16; (b) PR with code length 32; (c) PR with code length 64; (d) PR with code length 128.
Figure 5. PR on CIFAR-10. (a) PR with code length 16; (b) PR with code length 32; (c) PR with code length 64; (d) PR with code length 128.
Symmetry 12 01193 g005
Figure 6. P@N on CIFAR-10. (a) P@N with code length 16; (b) P@N with code length 32; (c) P@N with code length 64; (d) P@N with code length 128.
Figure 6. P@N on CIFAR-10. (a) P@N with code length 16; (b) P@N with code length 32; (c) P@N with code length 64; (d) P@N with code length 128.
Symmetry 12 01193 g006
Figure 7. PR on NUS-WIDE. (a) PR with code length 16; (b) PR with code length 32; (c) PR with code length 64; (d) PR with code length 128.
Figure 7. PR on NUS-WIDE. (a) PR with code length 16; (b) PR with code length 32; (c) PR with code length 64; (d) PR with code length 128.
Symmetry 12 01193 g007
Figure 8. P@N on NUS-WIDE. (a) P@N with code length 16; (b) P@N with code length 32; (c) P@N with code length 64; (d) P@N with code length 128.
Figure 8. P@N on NUS-WIDE. (a) P@N with code length 16; (b) P@N with code length 32; (c) P@N with code length 64; (d) P@N with code length 128.
Symmetry 12 01193 g008
Table 1. Configuration information.
Table 1. Configuration information.
ItemConfiguration
OS
GPU
Ubuntu 16.04 (X64)
4*GTX 1080ti
Table 2. Mean Average Precision (MAP) of different η .
Table 2. Mean Average Precision (MAP) of different η .
η CIFAR-10 (MAP@ALL) (bits)NUS-WIDE (MAP@5000) (bits)
163264128163264128
00.10150.30910.37440.37390.65710.69760.72300.7409
0.10.10150.30800.34130.36680.65460.69850.71980.7356
0.20.29400.32040.34640.37040.64800.70070.71830.7340
0.30.28410.32830.35080.36880.64440.69430.72130.7343
0.40.28540.33090.3470 0.36530.64340.68920.71790.7318
0.50.28370.31950.34040.36290.62350.67680.71310.6979
0.60.28220.30960.32760.33470.61990.67050.6940 0.6892
0.70.21990.24280.2650 0.29930.57610.62940.66610.7120
0.80.16770.20080.2380 0.2900 0.52390.57950.62820.6979
0.90.15760.19320.23340.28880.49200.55220.61680.6926
Table 3. MAP for Different Number of Bits.
Table 3. MAP for Different Number of Bits.
MethodCIFAR-10 (MAP @ALL) (bits)NUS-WIDE (MAP @5000) (bits)
163264128163264128
SSDH [32]0.21720.23130.24050.23510.58600.57880.60790.6127
SSDH+GA0.28270.30180.32190.34620.65280.66370.67870.7032
Ours0.27040.31330.32400.34520.65120.68460.70390.7125
Ours+GA0.29360.32330.34990.36610.66010.69810.72240.7359
Table 4. MAP for different numbers of bits.
Table 4. MAP for different numbers of bits.
MethodCIFAR-10 (MAP @ALL) (bits)NUS-WIDE (MAP @5000) (bits)
163264128163264128
SH [37]0.16210.15750.15490.15480.43500.41290.40420.4100
SPH [38]0.16640.17450.19310.19920.44580.45370.49260.5.00
ITQ [13]0.19990.20980.22290.23120.50820.51390.52520.5311
DSH [39]0.16860.18460.19550.20710.51230.51180.51100.5267
SSDH [32]0.21720.23130.24050.23510.58600.57880.60790.6127
HashGAN [34]0.24570.26780.30770.34220.40200.42070.44600.4883
USDH [31]0.26130.30210.31220.32410.64070.65680.65870.6724
Ours+GA0.29360.32330.34990.36610.66010.69810.72240.7359

Share and Cite

MDPI and ACS Style

Jiang, S.; Wang, L.; Cheng, S.; Du, A.; Li, Y. Unsupervised Hashing with Gradient Attention. Symmetry 2020, 12, 1193. https://doi.org/10.3390/sym12071193

AMA Style

Jiang S, Wang L, Cheng S, Du A, Li Y. Unsupervised Hashing with Gradient Attention. Symmetry. 2020; 12(7):1193. https://doi.org/10.3390/sym12071193

Chicago/Turabian Style

Jiang, Shaochen, Liejun Wang, Shuli Cheng, Anyu Du, and Yongming Li. 2020. "Unsupervised Hashing with Gradient Attention" Symmetry 12, no. 7: 1193. https://doi.org/10.3390/sym12071193

APA Style

Jiang, S., Wang, L., Cheng, S., Du, A., & Li, Y. (2020). Unsupervised Hashing with Gradient Attention. Symmetry, 12(7), 1193. https://doi.org/10.3390/sym12071193

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop