학술논문

Large Scale Near-Duplicate Image Retrieval via Patch Embedding
Document Type
Conference
Source
2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) ICCVW Computer Vision Workshop (ICCVW), 2019 IEEE/CVF International Conference on. :2972-2979 Oct, 2019
Subject
Computing and Processing
Visualization
Training
Image retrieval
Feature extraction
Indexes
Distortion
Task analysis
NDIR
Bag of Words
Feature Match
Neural Networks
Patch Embedding
Language
ISSN
2473-9944
Abstract
Large scale near-duplicate image retrieval (NDIR) relies on the Bag-of-Words methodology which quantizes local features into visual words. However, the direct match of these visual words typically leads to unpleasant mismatches due to quantization errors. To enhance the discriminability of the matching process, existing methods usually exploit hand-crafted contextual information, which has limited performance in complicated real-world scenarios. In contrast, we in this paper propose a trainable lightweight embedding network to extract local binary features. The network takes image patches as inputs and generates the binary code that can be efficiently stored in the inverted indexing file and helps discard mismatches immediately during the retrieval process. We improve the discriminability of the code by elaborately composing the training patches for network optimization, which consists of a proper inter-class (non-duplicate) patches selection and a rich intra-class (near-duplicate) patch generation. We evaluate our approach on the open NDIR dataset, INRIA CopyDays, and the experimental results show that our method performs favorably against the state-of-the-art algorithms. Furthermore, with a relatively short code length, our approach achieves higher query speed and lower storage occupation.