Training deep retrieval models with noisy datasets: Bag exponential loss Articles uri icon

publication date

  • April 2021

start page

  • 1

end page

  • 13

volume

  • 112

International Standard Serial Number (ISSN)

  • 0031-3203

Electronic International Standard Serial Number (EISSN)

  • 1873-5142

abstract

  • Although the CNNs are a very powerful tool for image retrieval, the need of training datasets properly adapted to the application at hand hinders the usefulness of such networks, specially since the datasets need to be free of noise to avoid spoiling the learning process. An ad hoc preprocessing of the dataset to mitigate the noise is a possible solution, but it is usually non-trivial and requires significant human intervention. In this paper, we pave the road for training CNNs for image retrieval with noisy datasets. In particular, we propose a novel Bag Exponential Loss function that, inspired by the Multiple Instance Learning framework, works with bags of matching images instead of single pairs, and allows a dynamical weighting of the relevance of each sample as the training progresses. The formulation of the proposed model is general enough and may serve to other purposes than dealing with noise if parameters are chosen appropriately. Extensive experimental results show the superior performance of the proposed loss with respect to the current state-of-the-art as well as its ability to cope with noisy training sets. Pytorch code available in ttps://github.com/tmcortes/BELoss

subjects

  • Telecommunications

keywords

  • image retrieval; noise; multiple instance learning; loss functions