Rethinking Neural Networks with Benford’s Law

Neurips 2021, Machine Learning for Physics, 2021

Recommended citation: Sahu, S. K., Java, A., Shaikh, A., & Kilcher, Y. (2021). Rethinking Neural Networks With Benford's Law. arXiv preprint arXiv:2102.03313. https://ml4physicalsciences.github.io/2021/files/NeurIPS_ML4PS_2021_99.pdf

In this work, we address the following question: Is the distribution of the Neural Network parameters related to the network’s generalization capability? To that end, we first define a metric, MLH (Model Enthalpy), that measures the closeness of a set of numbers to Benford’s Law. Second, we use MLH as an alternative to Validation Accuracy for Early Stopping and provide experimental evidence that even if the optimal size of the validation set is known beforehand, the peak test accuracy attained is lower than early stopping with MLH i.e. not using a validation set at all. This work was done with my research group.