Jason Wei
@_jasonwei
Followers
102K
Following
10K
Media
130
Statuses
1K
ai researcher @meta superintelligence labs, past: openai, google đ§
sf
Joined October 2020
Kickstarting my career in comedy after all this AGI stuff is over. And yes, this was actually the first time Sam, Hyung Won, and Max heard the joke
79
69
2K
Bullish, in the coming decades majority of compute will be spent on ai for science
Today, @ekindogus and I are excited to introduce @periodiclabs. Our goal is to create an AI scientist. Science works by conjecturing how the world might be, running experiments, and learning from the results. Intelligence is necessary, but not sufficient. New knowledge is
13
18
406
Excited to share that I recently joined the MSL team! Building personal superintelligence is serious and fun here. Join us!
After a great time at OpenAI, we (@EdwardSun0909, @_jasonwei) recently joined @Meta Superintelligence Labs. The first month has already been so much fun building from a clean slate with a truly talent-dense team! Very excited about the compute and long term focus of the new lab
57
26
852
Old friends, new lab
After a great time at OpenAI, we (@EdwardSun0909, @_jasonwei) recently joined @Meta Superintelligence Labs. The first month has already been so much fun building from a clean slate with a truly talent-dense team! Very excited about the compute and long term focus of the new lab
60
31
985
Becoming an RL diehard in the past year and thinking about RL for most of my waking hours inadvertently taught me an important lesson about how to live my own life. One of the big concepts in RL is that you always want to be âon-policyâ: instead of mimicking other peopleâs
127
347
3K
New blog post about asymmetry of verification and "verifier's law": https://t.co/bvS8HrX1jP Asymmetry of verificationâthe idea that some tasks are much easier to verify than to solveâis becoming an important idea as we have RL that finally works generally. Great examples of
54
247
2K
Bryan Johnson longevity mix is the most popular drink at ragers in SF
16
3
201
We donât have AI self-improves yet, and when we do it will be a game-changer. With more wisdom now compared to the GPT-4 days, it's obvious that it will not be a âfast takeoffâ, but rather extremely gradual across many years, probably a decade. The first thing to know is that
84
169
1K
The most rewarding thing about working in the office on nights and weekends is not the actual work you get done, but the spontaneous conversations with other people who are always working. Theyâre the people who tend to do big things and will become your most successful friends
30
31
811
I would say that we are undoubtedly at AGI when AI can create a real, living unicorn. And no I donât mean a $1B company you nerds, I mean a literal pink horse with a spiral horn. A paragon of scientific advancement in genetic engineering and cell programming. The stuff of
80
42
749
The greatest contribution of human language is bootstrapping language model training
8
8
182
AI research is strange in that you spend a massive amount of compute on experiments to learn simple ideas that can be expressed in just a few sentences. Literally things like âtraining on A generalizes if you add Bâ, âX is a good way to design rewardsâ, or âthe fact that method M
25
38
601
My favorite thing an old OpenAI buddy of mine told me is, whenever he hears that someone is a âgreat AI researcherâ, he just directly spends 5 minutes looking at that personâs PRs and wandb runs. People can do all kinds of politics and optical shenanigans, but at the end of the
37
64
992
One way of thinking about what AI will automate first is via the âdescription-execution gapâ: how much harder is it to describe the task than to actually do it? Tasks with large description-execution gaps will be ripe for automation because itâs easy to create training data and
17
47
391
RL environment specs are among the most consequential things we can write as AI researchers. A relatively short spec (e.g., <1000 words of instructions saying what problems to create and how to grade them) often gets expanded either by humans or via synthetic methods into
13
35
445
Itâs actually a good thing these days to have subtle grammar errors in your writing. It sprinkles on a clear human touch. You never want your reader questioning if what theyâre reading was written or edited by chatgtp
67
13
408
Was attending a talk in a big lecture hall and the guy in front of me had the craziest conversation with ChatGPT for the whole hour about how to get his girlfriend back. Dozens of messages of pasting screenshots of text conversations to analyze tone of responses; whether to
13
8
421
OK as someone pointed out ChatGPT using nothing from chatbot research isn't totally accurate. What I meant to say is that much of chatbot research that was mainstream at some point in time (e.g., dialogue state tracking, or slot filling, or semantic parsing) wasn't used in
2
3
25
The 80-20 rule happens often in AI research, where you get 80% of the payoff from the first 20% of the effort. But there is often also an inverse rule, where itâs actually the final 20% of that yields 80% of the payout. Some examples: 1. When your eval is already good in many
13
27
383
There are traditionally two types of research: problem-driven research and method-driven research. As weâve seen with large language models and now AlphaEvolve, it should be very clear now that total method-driven research is a huge opportunity. Problem-driven research is nice
21
92
716