Ambiguous out-of-distribution generalization on an algorithmic task
Introduction It's now well known that simple neural network models often "grok" algorithmic tasks. That is, when trained for many epochs on a subset of the full input space, the model quickly attains perfect train accuracy and then, much later, near-perfect test accuracy. In the former phase, the model memorizes...
Great to see more work on (better) influence functions!
Lots of interesting things to discuss here[1], but one thing I would like to highlight is that classical IFs indeed arise when you do the usual implicit function theorem + global minimum assumption (which is obviously violated in the context of DL), but they also arise as the limit of unrolling as t→∞. What follows will be more of theoretical nature summarizing statements in Mlodozeniec et al.
Influence functions suffer from another shortcoming, since they only use final weights (as you are aware). So you might say that we shouldn't do influence functions, but track a different counterfactual: The counterfactual over training "What if I added/removed... (read 351 more words →)