Authors
Nicolas Lell, Ansgar Scherp
Publication date
2023/8/22
Book
International Cross-Domain Conference for Machine Learning and Knowledge Extraction
Pages
200-226
Publisher
Springer Nature Switzerland
Description
When training a Neural Network, it is optimized using the available training data with the hope that it generalizes well to new or unseen testing data. At the same absolute value, a flat minimum in the loss landscape is presumed to generalize better than a sharp minimum. Methods for determining flat minima have been mostly researched for independent and identically distributed (i.i.d.) data such as images. Graphs are inherently non-i.i.d. since the vertices are edge-connected. We investigate flat minima methods and combinations of those methods for training graph neural networks (GNNs). We use GCN and GAT as well as extend Graph-MLP to work with more layers and larger graphs. We conduct experiments on small and large citation, co-purchase, and protein datasets with different train-test splits in both the transductive and inductive training procedure. Results show that flat minima methods can improve the …
Total citations
Scholar articles
N Lell, A Scherp - International Cross-Domain Conference for Machine …, 2023