~2yrs ago
@nsaphra
came to my poster & we discussed regularizing to ctrl interpretability. She mentioned a superstar grad student (
@_angie_chen
). Things really got wild when
@ziv_ravid
joined the party. And
@kchonyc
graced us w/ wisdom throughout. V excited to finally announce:
New work w/
@ziv_ravid
@kchonyc
@leavittron
@nsaphra
: We break the steepest MLM loss drop into *2* phase changes: first in internal grammatical structure, then external capabilities. Big implications for emergence, simplicity bias, and interpretability! ๐งต