Show simple item record

dc.contributor.authorDavel, Marelie Hattingh
dc.contributor.authorBarnard, Etienne
dc.contributor.authorTheunissen, Marthinus Wilhelmus
dc.date.accessioned2021-03-18T07:21:30Z
dc.date.available2021-03-18T07:21:30Z
dc.date.issued2020
dc.identifier.citationMarthinus Wilhelmus Theunissen, Marelie H. Davel and Etienne Barnard, "Benign interpolation of noise in deep learning", South African Computer Journal, Vol 32, no 2, pp 80-101, December 2020en_US
dc.identifier.issn1015-7999
dc.identifier.issn2313-7835
dc.identifier.urihttp://hdl.handle.net/10394/36915
dc.description.abstractThe understanding of generalisation in machine learning is in a state of flux, in part due to the ability of deep learning models to interpolate noisy training data and still perform appropriately on out-of-sample data, thereby contradicting long-held intuitions about the bias-variance trade off in learning. We expand upon relevant existing work by discussing local attributes of neural network training within the context of a relatively simple framework.We describe how various types of noise can be compensated for within the proposed framework in order to allow the deep learning model to generalise in spite of interpolating spurious function descriptors. Empirically,we support our postulates with experiments involving overparameterised multilayer perceptrons and controlled training data noise. The main insights are that deep learning models are optimised for training data modularly, with different regions in the function space dedicated to fitting distinct types of sample information. Additionally,we show that models tend to fit uncorrupted samples first. Based on this finding, we propose a conjecture to explain an observed instance of the epoch-wise double-descent phenomenon. Our findings suggest that the notion of model capacity needs to be modified to consider the distributed way training data is fitted across sub-units.en_US
dc.language.isoenen_US
dc.publisherSouth African Institute of Computer Scientists and Information Technologistsen_US
dc.subjectDeep Learningen_US
dc.subjectMachine Learningen_US
dc.subjectLearning Theoryen_US
dc.subjectGeneralizationen_US
dc.titleBenign interpolation of noise in deep learningen_US
dc.typeArticleen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record