Posts

Stan van Wingerden's Shortform 2024-12-11T18:00:53.568Z
Timaeus is hiring! 2024-07-12T23:42:28.651Z
Timaeus's First Four Months 2024-02-28T17:01:53.437Z
Announcing Timaeus 2023-10-22T11:59:03.938Z
You’re Measuring Model Complexity Wrong 2023-10-11T11:46:12.466Z
Open Call for Research Assistants in Developmental Interpretability 2023-08-30T09:02:59.781Z
Apply for the 2023 Developmental Interpretability Conference! 2023-08-25T07:12:36.097Z
Towards Developmental Interpretability 2023-07-12T19:33:44.788Z

Comments

Comment by Stan van Wingerden (stan-van-wingerden) on Stan van Wingerden's Shortform · 2024-12-12T16:48:27.225Z · LW · GW

Here's the plot, which is very similar to Experience Machine's:

My conclusion from this is that the LLC and the L2 norm measure basically the same thing in this setup. They don't always: for further comparison with more unprincipled metrics in more complex setups, see comparisons with weight norm / Hessians in fig 22, 23, and 25 here and comparisons with Hessian-based methods and ablations here

Comment by Stan van Wingerden (stan-van-wingerden) on Stan van Wingerden's Shortform · 2024-12-11T18:00:53.900Z · LW · GW

Here's a quick interesting-seeming devinterp result:

We can estimate the Local Learning Coefficient (LLC, the central quantity of Singular learning theory, for more info see these posts / papers) of a simple grokking model on its training data over the course of training. 

This yields the following plot:

(note: estimated LLC = lambdahat = )

What's interesting about this is that the estimated LLC of the model in this plot closely tracks test loss, even though it is estimated on training data.

On the one hand this is unsurprising: SLT predicts that the LLC determines the Bayes generalization error in the Bayesian setting.[1] On the other hand this is quite surprising: the Bayesian setting is not the same as SGD, an increase in training steps is not the same as an increase in the total number of samples, and the Bayes generalization is not exactly the same as test loss. Despite these differences, the LLC clearly tracks (in-distribution) generalization here. We see this as a positive sign for applying SLT to study neural networks trained by SGD.

This plot was made using the devinterp  python package, and the code to reproduce it (including hyperparameter selection) is available as a notebook at https://github.com/timaeus-research/devinterp/blob/main/examples/grokking.ipynb

Thanks to Nina Panickserry and Dmitry Vaintrob, whose earlier post on learning coefficients of modular addition served as the basis for this experiment.

 

  1. ^

    More precisely: in the Bayesian setting the Bayes generalization error, as a function of the number of samples n, is λ/n in leading order.

Comment by Stan van Wingerden (stan-van-wingerden) on Open Call for Research Assistants in Developmental Interpretability · 2023-08-30T09:45:02.132Z · LW · GW

It's in USD (should be reflected in to the announcement now)

Comment by Stan van Wingerden (stan-van-wingerden) on Crystal Healing — or the Origins of Expected Utility Maximizers · 2023-06-30T19:11:47.987Z · LW · GW

This last paragraph will live in my head forever, but I'm confused how it relates to the rest of the post. Would you agree with the following rephrasing? "Forming incomplete preferences (and thus not optimizing fully) is the easy way out, as it avoids taking some sure losses. But in doing so it also necessarily loses out on sure gains."