Today’s session brings Influence Functions under the spotlight - the theory, non-convergence issues, and uses for data pruning. Fabio will uncover the fragile nature of influence functions in deep learning, helping us understand what neural networks memorize, and exploring the possibility of beating power law scaling of model performance with dataset size.
Influence functions and Data Pruning: from theory to non-convergence
References
[Koh17U]
Understanding Black-box Predictions via Influence Functions,
[Bas20I]
Influence Functions in Deep Learning Are Fragile,
[Fel20W]
What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation,
[Sor22N]
Beyond neural scaling laws: beating power law scaling via data pruning,