Zhang, JingDai, YuchaoZhang, TongHarandi, MehrtashBarnes, NickHartley, Richard2021-07-312021-07-312021-07-312021-08-0110.1109/TPAMI.2020.3046486https://infoscience.epfl.ch/handle/20.500.14299/180348WOS:000670578800026The advances made in predicting visual saliency using deep neural networks come at the expense of collecting large-scale annotated data. However, pixel-wise annotation is labor-intensive and overwhelming. In this paper, we propose to learn saliency prediction from a single noisy labelling, which is easy to obtain (e.g., from imperfect human annotation or from unsupervised saliency prediction methods). With this goal, we address a natural question: Can we learn saliency prediction while identifying clean labels in a unified framework? To answer this question, we call on the theory of robust model fitting and formulate deep saliency prediction from a single noisy labelling as robust network learning and exploit model consistency across iterations to identify inliers and outliers (i.e., noisy labels). Extensive experiments on different benchmark datasets demonstrate the superiority of our proposed framework, which can learn comparable saliency prediction with state-of-the-art fully supervised saliency methods. Furthermore, we show that simply by treating ground truth annotations as noisy labelling, our framework achieves tangible improvements over state-of-the-art methods.Computer Science, Artificial IntelligenceEngineering, Electrical & ElectronicComputer ScienceEngineeringnoise measurementlabelingpredictive modelsannotationstrainingtask analysissaliency detectionsalinecy predictionsingle noisy labellingrobust model fittingobject detectionLearning Saliency From Single Noisy Labelling: A Robust Model Fitting Perspectivetext::journal::journal article::research article