NWU Institutional Repository

Exploring neural network training dynamics through binary node activations

dc.contributor.authorHaasbroek, Daniël G.
dc.contributor.authorDavel, Marelie H.
dc.date.accessioned2021-02-26T14:01:08Z
dc.date.available2021-02-26T14:01:08Z
dc.date.issued2020
dc.description.abstractEach node in a neural network is trained to activate for a specific region in the input domain. Any training samples that fall within this domain are therefore implicitly clustered together. Recent work has highlighted the importance of these clusters during the training process but has not yet investigated their evolution during training. Towards this goal, we train several ReLU-activated MLPs on a simple classification task (MNIST) and show that a consistent training process emerges: (1) sample clusters initially increase in size and then decrease as training progresses, (2) the size of sample clusters in the first layer decreases more rapidly than in deeper layers, (3) binary node activations, especially of nodes in deeper layers, become more sensitive to class membership as training progresses, (4) individual nodes remain poor predictors of class membership, even if accurate when applied as a group. We report on the detail of these findings and interpret them from the perspective of a high-dimensional clustering process.en_US
dc.identifier.isbn978-0-620-89373-2
dc.identifier.urihttp://hdl.handle.net/10394/36796
dc.language.isoenen_US
dc.publisherSouthern African Conference for Artificial Intelligence Researchen_US
dc.subjectNeural networksen_US
dc.subjectGeneralizationen_US
dc.subjectClusteringen_US
dc.titleExploring neural network training dynamics through binary node activationsen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Haasbroek-2020-exploring-nn-training.pdf
Size:
1.53 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.61 KB
Format:
Item-specific license agreed upon to submission
Description: