robertzzk Profile Banner
Rob K Profile
Rob K

@robertzzk

Followers
452
Following
2K
Media
301
Statuses
3K

Focused on AI work.

New York City
Joined October 2014
Don't wanna be here? Send us removal request.
@robertzzk
Rob K
4 months
RT @IvanArcus: 🧵Chain-of-Thought reasoning in LLMs like Claude 3.7 and R1 is behind many recent breakthroughs. But does the CoT always expl….
0
64
0
@robertzzk
Rob K
9 months
RT @peterwildeford: 🚨 IAPS is hiring 🚨. We seek Researchers / Senior Researchers to join our team to identifies concrete interventions that….
0
14
0
@robertzzk
Rob K
1 year
Outer loss transmits very limited bits of information through the info-theoretic bottleneck of the training process. so yes, if your function family (eg wetware brains) priors mesa-optimizers relatively densely, no sparsely delivered outer loss (death in the EEA) suffices. .
@ESYudkowsky
Eliezer Yudkowsky ⏹️
2 years
"Train it to be nice" is the obvious thought. Alas, I predict that one idiom that does generalize from natural selection to gradient descent, is that training on an outer loss gets you something not internally aligned to that outer loss. It gets you ice cream and condoms.
0
0
0
@robertzzk
Rob K
1 year
Disagree. "Gradient-descenting" is the wrong verb as it implies attribution to kindness-likelihood is primarily the optimizer. The text that is to be predicted matters; the etiology of the ontology implicit in the training distribution *was* shaped by m-s of proteins and IS kind.
@ESYudkowsky
Eliezer Yudkowsky ⏹️
2 years
Gradient-descending an AI system to predict text, or even to play video games, is nothing like this. It is exploring nowhere near this space. Gradient descent of matrices is not mutation-selection of proteins and I don't expect it to hit on anything like similar architectures.
1
0
1
@robertzzk
Rob K
1 year
RT @akothari: The Microsoft / CrowdStrike outage has taken down most airports in India. I got my first hand-written boarding pass today 😅 h….
0
11K
0
@robertzzk
Rob K
1 year
RT @Eth_Toripo: @spectatorindex Amsterdam Airport atm
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
118
0
@robertzzk
Rob K
1 year
What does fine tuning do to models? Do representations transfer between base and chat versions? Find out in the next episode of @Connor_Kissane and Rob project!.
@Connor_Kissane
Connor Kissane
1 year
New post with @robertzzk, @ArthurConmy, & @NeelNanda5: Sparse Autoencoders (usually) Transfer between Base and Chat Models! This suggests that models' representations remain extremely similar after fine-tuning.
Tweet media one
0
0
1
@robertzzk
Rob K
1 year
RT @Connor_Kissane: New post with @robertzzk, @ArthurConmy, & @NeelNanda5: Sparse Autoencoders (usually) Transfer between Base and Chat Mod….
0
4
0
@robertzzk
Rob K
1 year
Open X for *one* moment and then it's, like, 3 hours of ML paper reading later. 👋 cya crazy peeps.
0
0
2
@robertzzk
Rob K
1 year
My feed is nothing except COLM paper acceptances. I get, it you got in, nice job. I'm very proud of you.
0
0
2
@robertzzk
Rob K
1 year
Sparse Autoencoders help us understand the MLPs of LLMs, but what's up with attention?. Find out in our new paper with @Connor_Kissane and @NeelNanda5!.
@Connor_Kissane
Connor Kissane
1 year
Sparse Autoencoders help us understand the MLPs of LLMs, but what's up with attention?. In our new paper with @NeelNanda5, we introduce Attention Output SAEs to uncover what concepts attention layers learn. Further, we use them to find novel insights previously out-of-reach!🧵
Tweet media one
0
0
4
@robertzzk
Rob K
1 year
One step closer to a full symbolic notation for dance:
0
0
0
@robertzzk
Rob K
1 year
New attention sparseautoencoders post dropped. $1000 bounty to whomever find the best attention circuit!.
@NeelNanda5
Neel Nanda
1 year
Great post from my scholars @Connor_Kissane.& @robertzzk!. SAEs are fashionable, but are they a useful tool for researchers? They are! We find a deeper understanding of the well-studied IOI circuit, and make a circuit analysis tool. $1000 bounty to whoever finds the best circuit!.
0
1
5
@robertzzk
Rob K
1 year
Now, Step 3.
0
0
0
@robertzzk
Rob K
2 years
RT @robertghilduta: Anyone know of any publications or research on changing out the activation function of trained transformer network? Loo….
0
1
0
@robertzzk
Rob K
3 years
RT @chmonke:
Tweet media one
0
47
0
@robertzzk
Rob K
3 years
That’s just crazy talk.
@idavidrein
david rein
3 years
What if we like, didn't train AI to intentionally lie and deceive??.
0
0
3
@robertzzk
Rob K
3 years
At hotpot. Don’t know what omasum is. Google it. “The omasum, also known as the bible, the fardel, the manyplies, the psalterium, is the 3rd compartment of the stomach in ruminants. The omasum comes after rumen & reticulum, before abomasum.” the rick/morty altverse is leaking 😱.
0
0
0
@robertzzk
Rob K
3 years
Guy who’s freaking out because he uses past tweets as knowledge mgmt tool.
2
1
10
@robertzzk
Rob K
3 years
Happy Monday fellow 🐦 people. First week with our brave new 🐦 dad, 🍈 Musk.
0
0
2