日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 >

Bit-Scalable Deep Hashing with Regularized Similarity Learning for Image Retrieval and Person Re-ide

發布時間:2025/4/16 32 豆豆
生活随笔 收集整理的這篇文章主要介紹了 Bit-Scalable Deep Hashing with Regularized Similarity Learning for Image Retrieval and Person Re-ide 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

總結:(1)對基于三元組學習的損失函數利用拉普拉斯矩陣增加了規則化項
(2)通過給hashing code 的每一Bit 給定一個權重,從而實現Bit-Scalable
(3)設計tanh-like layer
(4)設計基于圖像的優化算法

Hash—> fast approximate similarity search

Abstract

Two crucial steps in image retrieval:

(1)Extracting informative image features
(2)Learning effective approximate hashing function

Conventional methods:

(1)Learn hash functions from a predefined hand-craft feature space
(2)Preset the bit lengths of output hashing codes

This paper:

we propose a supervised learning framework to generate compact and bit-scalable hashing
codes directly from raw images.—>Pose hashing learning as a problem of regularized similarity learning.
? Organize the training images to triplet samples
? Maximize the margin
? Regularization term –>adjacency consistency
? Unequally weights for bits of hashing codes.

Experiments:

Outperforms state-of-arts on public benchmarks of image retrieval.
Shows promising results in application of Person re-identification.
Bit-scalable hasing codes preserve the discriminative powers with shorter code lengths.

Introduction

(1) Traditional methods:
The feature representation may not be tailored to the objective of hashing learning.
Moreover, the hand-crafted feature engineering often requires much domain knowledge and heavy tuning.
(2)Most approaches–>preset lengths
However, one often requires hashing codes of different lengths under different scenarios.
E.g. ,mobile devices
To cope with such requirements:
one solution is store several versions of hashing codes in different lengths.–>extra computation and storage.
Bit-scalable hashing methods: generate hashing codes bit by bit in a significance decent way.–> need carefully design the embedded feature space
the performances may dramatically fall when shortening the hashing codes.

This paper:
Supervised Bit-Scalable Deep Hashing framework:
(1)CNN
(2)Each bit of hashing codes is weighted according to their significance

Main components:

(1)Present a novel formulation of relative similarity comparison based on the triplet-based model.—–>
we extend the triplet-based relative comparison by incorporating a regularization term, partially motivated by the recently proposed Laplacian Sparse Coding.
(2)Adopt the deep CNN architecture to extract the discriminative features from the input images.
One FC layer and one tanh-like layer—->output the binary hashing codes
An element-wise layer–>to weight the bit of hashing code.
(3)Implements the learning algorithm in a batch-process fashion.
Organize the triplet samples from a randomly selected subset ( 150~200) of the training images.
Use stochastic gradient decent (SGD) for parameter learning.
(Calculate the partial derivative on images instead of on triplets samples.—->reduce the computational cost which is linear to the selected subset of images.)

Contributions

(1) It uni?es feature learning and hash function learning via deep neural networks, and the proposed bit-scalable hashing learning can effectively improves the ?exibility of image retrieval.
(2) It presents a novel formulation (i.e., the regularized triplet-based comparison) for Hashing learning.
(3)Extensive experiments on standard benchmarks demonstrate:
The learned hashing codes well preserve instance-level similarity
Outperforms state-of-the-art hashing learning approaches.
Successfully apply the method to person re-identification.

Hashing methods: data-independent and data-dependent.
Data-independent: without any training , make the codes scattered
Data-dependent: How to learn compact hashing codes from the training data .
(1)projecting the high dimensional features onto the lower dimensional space.
(2)quantizing the real-valued representation into binary codes.
Deep learning methods:
[2] proposed a learning-to-rank framework based on multi-scale neural networks, and showed promising performance on capturing fine-grained image similarity.
Pre-training on ImageNet.
[35] utilized CNN for supervised hashing.
Produced the hashing codes by decomposing the pairwise similarity matrix, then learned the mapping fucntions from images to the codes.—>may fail to deal with large-scale data due to the matrix decomposition operation.

Experiments

Datasets : MNIST , CIFAR-10 , CIFAR-20 and NUS-WIDE
Split images into a training set and a query set
Leave -one-out
CHUK03
Several variants of the framework:
(1) without element-wise layer
DRSCH (Deep Regularized Similarity Comparison Hashing)
DSCH
To verify the effectiveness of the regularization term.
(2)remove tanh-like layer —>evaluate the effectiveness of tanh-like
Euclidean
(3)Bit-scale
BS-DRSCH
BS-DSCH

Four types of evaluation tools:
mAP: http://www.zhihu.com/question/41540197
precision@500
precision@k
HAM2

總結

以上是生活随笔為你收集整理的Bit-Scalable Deep Hashing with Regularized Similarity Learning for Image Retrieval and Person Re-ide的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。