Seungwook Han Profile
Seungwook Han

@seungwookh

Followers
404
Following
638
Media
21
Statuses
135

phd-ing @MIT_CSAIL, prev @MITIBMLab @columbia

Joined June 2017
Don't wanna be here? Send us removal request.
@seungwookh
Seungwook Han
6 months
🧙‍♂️Excited to share our new whitepaper “General Reasoning Requires Learning to Reason from the Get-Go.” . We argue that simply making models bigger and feeding them more data is NOT enough for robust, adaptable reasoning. (1/n).
1
12
79
@seungwookh
Seungwook Han
14 days
RT @jyo_pari: We have a fun collaboration of @GPU_MODE x @scaleml coming up!. We’re hosting a week-long online bootcamp that explores the c….
0
20
0
@grok
Grok
19 days
Blazing-fast image creation – using just your voice. Try Grok Imagine.
283
567
3K
@seungwookh
Seungwook Han
1 month
to clarify, not saying we’re there atm. i dont have a formal definition of what it means to be human, but agency and the ability to continually learn seem to be important.
0
0
0
@seungwookh
Seungwook Han
1 month
this tendency to anthropomorphize is too real and shuns me from reading work with such titles. on the other hand, a part of me also asks: how do we know when a thing is conscious and start to analyze as if it is another human-like organism. we automatically assume all humans are.
@fchollet
François Chollet
1 month
Resist the tendency to anthropomorphize that which is not human.
1
0
3
@seungwookh
Seungwook Han
1 month
uncertainty-aware reasoning, akin to how humans leverage our confidence.
@MehulDamani2
Mehul Damani
1 month
🚨New Paper!🚨.We trained reasoning LLMs to reason about what they don't know. o1-style reasoning training improves accuracy but produces overconfident models that hallucinate more. Meet RLCR: a simple RL method that trains LLMs to reason and reflect on their uncertainty --
Tweet media one
0
1
3
@seungwookh
Seungwook Han
1 month
was actually wondering with @hyundongleee the fundamental differences between diffusion and autoregressive modeling other than the structure imposed in the modeling of the sequential conditional distribution and how they manifest. a poignant paper that addresses this thought.
@mihirp98
Mihir Prabhudesai
1 month
🚨 The era of infinite internet data is ending, So we ask:. 👉 What’s the right generative modelling objective when data—not compute—is the bottleneck?. TL;DR:. ▶️Compute-constrained? Train Autoregressive models. ▶️Data-constrained? Train Diffusion models. Get ready for 🤿 1/n
Tweet media one
0
1
13
@seungwookh
Seungwook Han
1 month
omw to trying this out đź‘€.
@pika_labs
Pika
1 month
Some news: We're building the next big thing — the first-ever AI-only social video app, built on a highly expressive human video model. Over the past few weeks, we’ve been testing it in private beta. Now, we’re opening early access: download the iOS app to join the waitlist, or.
0
0
0
@seungwookh
Seungwook Han
1 month
how particles can act differently under different scales and conditions and how we can equip it as part of design is cool.
@MITarchitecture
MIT Architecture
8 months
8. Jeonghyun Yoon: Precisely Loose: Unraveling the Potential of Particles. A big thank you goes out to the entire architecture community, including advisors, readers, staff, family and peers who helped bring these projects to light. Image credit: Chenyue “xdd” Dai . 2/2
Tweet media one
0
0
4
@seungwookh
Seungwook Han
1 month
RT @LakerNewhouse: [1/9] We created a performant Lipschitz transformer by spectrally regulating the weights—without using activation stabil….
0
78
0
@seungwookh
Seungwook Han
2 months
But actually this is the og way of doing it and should stop by E-2103 to see @jxbz and Laker Newhouse whiteboard the whole paper.
Tweet media one
@jxbz
Jeremy Bernstein
2 months
Laker and I are presenting this work in an hour at ICML poster E-2103. It’s on a theoretical framework and language (modula) for optimizers that are fast (like Shampoo) and scalable (like muP). You can think of modula as Muon extended to general layer types and network topologies
Tweet media one
1
6
74
@seungwookh
Seungwook Han
2 months
RT @jyo_pari: If you are interested in questioning how we should pretrain models and create new architectures for general reasoning . - the….
0
6
0
@seungwookh
Seungwook Han
2 months
Presenting our ICML spotlight poster at today 11am @ E-606 w/ @jyo_pari!. We need to fundamentally change how we train to achieve true reasoning. Reward-based Pretraining (RPT) > Supervised Pretraining.
@seungwookh
Seungwook Han
6 months
🧙‍♂️Excited to share our new whitepaper “General Reasoning Requires Learning to Reason from the Get-Go.” . We argue that simply making models bigger and feeding them more data is NOT enough for robust, adaptable reasoning. (1/n).
1
2
14
@seungwookh
Seungwook Han
2 months
🤗
Tweet media one
2
0
8
@seungwookh
Seungwook Han
2 months
How do task vectors emerge during pretraining—and can they predict ICL performance?. Come see our ICML spotlight poster "Emergence and Effectiveness of Task Vectors in ICL" at 11am @ East Hall A-B (#E-2312) with @jinyeop_song!. 🔗
Tweet media one
1
5
13
@seungwookh
Seungwook Han
2 months
At #ICML 🇨🇦 this week. I'm convinced that the core computations are shared across modalities (vision, text, audio, etc). The real question is the (synthetic) generative process that ties them. Reach out if you have thoughts or want to chat!.
0
3
16
@seungwookh
Seungwook Han
2 months
wholeheartedly agree with this direction that games can be a good playground for learning reasoning. makes us think what other synthetic environments we can design and grow over complexity.
@Benjamin_eecs
Bo Liu (Benjamin Liu)
2 months
We've always been excited about self-play unlocking continuously improving agents. Our insight: RL selects generalizable CoT patterns from pretrained LLMs. Games provide perfect testing grounds with cheap, verifiable rewards. Self-play automatically discovers and reinforces
Tweet media one
0
1
7
@seungwookh
Seungwook Han
2 months
robot arms becoming more human-like. now with a wrist 🦾.
@martinpeticco
Martin Peticco
2 months
What’s keeping robot arms from working like human arms?. They're big, slow, have the wrong joints, and can't conform to their environment. DexWrist solves all of these issues and simplifies learning constrained, dynamic manipulation👉 
0
0
3
@seungwookh
Seungwook Han
3 months
RT @phillip_isola: Our computer vision textbook is now available for free online here:. We are working on adding so….
Tweet card summary image
visionbook.mit.edu
0
620
0
@seungwookh
Seungwook Han
3 months
a step towards self-learning models — self-synthesizing data to train on and evolving.
@jyo_pari
Jyo Pari
3 months
What if an LLM could update its own weights?. Meet SEAL🦭: a framework where LLMs generate their own training data (self-edits) to update their weights in response to new inputs. Self-editing is learned via RL, using the updated model’s downstream performance as reward.
Tweet media one
1
1
8