Discriminative Learning of Deep Convolutional Feature Point Descriptors

Deep learning has revolutionalized image-level tasks such as classification, but patch-level tasks, such as correspondence, still rely on hand-crafted features, e.g. SIFT. In this paper we use Convolutional Neural Networks (CNNs) to learn discriminant patch representations and in particular train a Siamese network with pairs of (non-)corresponding patches. We deal with the large number of potential pairs with the combination of a stochastic sampling of the training set and an aggressive mining strategy biased towards patches that are hard to classify. By using the L2 distance during both training and testing we develop 128-D descriptors whose euclidean distances reflect patch similarity, and which can be used as a drop-in replacement for any task involving SIFT. We demonstrate consistent performance gains over the state of the art, and generalize well against scaling and rotation, perspective transformation, non-rigid deformation, and illumination changes. Our descriptors are efficient to compute and amenable to modern GPUs, and are publicly available.

Published in:
Proceedings of the International Conference on Computer Vision
Presented at:
International Conference on Computer Vision, Santiago, Chile, December 13-16, 2015

 Record created 2015-10-27, last modified 2019-12-05

iccv-2015-deepdesc - Download fulltextPDF
iccv-2015-deepdesc-supp - Download fulltextPDF
Rate this document:

Rate this document:
(Not yet reviewed)