Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Disparity Between Batches as a Signal for Early Stopping
 
conference paper

Disparity Between Batches as a Signal for Early Stopping

Forouzesh, Mahsa  
•
Thiran, Patrick  
September 13, 2021
Machine Learning and Knowledge Discovery in Databases. Research Track
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2021)

We propose a metric for evaluating the generalization ability of deep neural networks trained with mini-batch gradient descent. Our metric, called gradient disparity, is the l2 norm distance between the gradient vectors of two mini-batches drawn from the training set. It is derived from a probabilistic upper bound on the difference between the classification errors over a given mini-batch, when the network is trained on this mini-batch and when the network is trained on another mini-batch of points sampled from the same dataset. We empirically show that gradient disparity is a very promising early-stopping criterion (i) when data is limited, as it uses all the samples for training and (ii) when available data has noisy labels, as it signals overfitting better than the validation data. Furthermore, we show in a wide range of experimental settings that gradient disparity is strongly related to the generalization error between the training and test sets, and that it is also very informative about the level of label noise.

  • Details
  • Metrics
Type
conference paper
DOI
10.1007/978-3-030-86520-7_14
Author(s)
Forouzesh, Mahsa  
•
Thiran, Patrick  
Date Issued

2021-09-13

Publisher

Springer

Published in
Machine Learning and Knowledge Discovery in Databases. Research Track
ISBN of the book

978-3-030865-20-7

978-3-030865-19-1

Series title/Series vol.

Lecture Notes in Computer Science; 12976

Subjects

Early Stopping

•

Generalization

•

Gradient Alignment

•

Overfitting

•

Neural Networks

•

Limited Datasets

•

Noisy Labels

Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
INDY2  
Event nameEvent placeEvent date
European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2021)

Bilbao, Basque Country, Spain

September 13-17, 2021

Available on Infoscience
September 17, 2021
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/181478
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés