dimension one
machine learning phenomena through minimalist examples
-
at interpolation, generalization is distance to Bayes
in the same setup as the previous post, the generalization gap turns out to have a simple analytical form: Bayes error plus a disagreement term, computable in one dimension
-
same model, same optimizer -- different generalization
a minimalist reproduction of why generalization cannot be understood without looking at the data