Abstract
This paper investigates sensitive minima in popular deep distance learning techniques such as Siamese and Triplet networks. We demonstrate that standard formulations may find solutions that are sensitive to small changes and thus do not generalize well. To alleviate sensitive minima we propose a new approach to regularize margin-based deep distance learning by introducing stochasticity in the loss that encourages robust solutions. Our experimental results on HPatches show promise compared to common regularization techniques including weight decay and dropout, especially for small sample sizes.
Original language | English |
---|---|
Article number | 9154359 |
Pages (from-to) | 145067-145076 |
Number of pages | 10 |
Journal | IEEE Access |
Volume | 8 |
DOIs | |
Publication status | Published - 2020 |
Keywords
- contrastive loss
- Deep metric learning
- feature point matching
- generalization
- regularization
- triplet loss