Sungmin Cha Profile
Sungmin Cha

@_sungmin_cha

Followers
463
Following
2K
Media
24
Statuses
432

Faculty Fellow @nyuniversity | PhD @SeoulNatlUni

Manhattan, NY
Joined July 2019
Don't wanna be here? Send us removal request.
@_sungmin_cha
Sungmin Cha
22 days
How can we be sure a generative model (LLMs, Diffusion) has truly unlearned something? What if existing evaluation metrics are misleading us? In our new paper, we introduce FADE, a new metric that assesses genuine unlearning by measuring distributional alignment, moving beyond
1
10
38
@NabeelSeedat01
Nabeel Seedat
18 hours
I'm excited to share recent research I've been working on with my amazing co-authors since joining @thomsonreuters Foundational Research, tackling two critical challenges in LLM evals: ๐—ฟ๐—ฒ๐—น๐—ถ๐—ฎ๐—ฏ๐—น๐—ฒ ๐—ฎ๐˜‚๐˜๐—ผ-๐—ฒ๐˜ƒ๐—ฎ๐—นs & ๐—ฐ๐—ผ๐—บ๐—ฝ๐˜‚๐˜๐—ฒ-๐—ฒ๐—ณ๐—ณ๐—ถ๐—ฐ๐—ถ๐—ฒ๐—ป๐˜ ๐—ฒ๐˜ƒ๐—ฎ๐—นs (Thread ๐Ÿ‘‡)
1
1
4
@ritaranx
Ran Xu
1 day
Happy to introduce my internship work at @Google and @GoogleDeepMind, collab w/ @googlecloud. We introduce TIR-Judge, an end-to-end agentic RL framework that trains LLM judges with tool-integrated reasoning ๐Ÿง ๐Ÿ› ๏ธ ๐Ÿ”— https://t.co/rtfqlvuzJ0 #Agents #LLMs #Judges #RL #reasoning
6
40
385
@lazyuniverse
Giri ATG
2 days
Launching our Research Lab : Advancing experience powered, decentralized superintelligence - built for continual learning, generalization & model-based planning. Press Release : https://t.co/iPYXb1nzYr Weโ€™re solving the hardest challenges in real-world industries, robotics,
Tweet card summary image
businesswire.com
ExperienceFlow.AI, a pioneer in delivering autonomous enterprise operations and decision-making platforms, announces launch of their Superintelligence Resear...
2
10
98
@dianarycai
Diana Cai
2 days
New changes for ICML 2026: - attendance not required for acceptance - original submission published along side camera ready version - new reciprocal reviewing requirements
@icmlconf
ICML Conference
2 days
- New guidelines on generative AI considerations Check out the full CfPs! Papers: https://t.co/4ppHEb6w1c Position Papers: https://t.co/HS6AXFehDW
5
6
119
@_sungmin_cha
Sungmin Cha
1 day
Letโ€™s go to Seoul!
@SharonYixuanLi
Sharon Li
2 days
๐Ÿ“ข We are excited to release the call for papers for #ICML2026, held in Seoul, South Korea next year! ๐Ÿ“… Key Dates Abstract deadline: Jan 23, 2026 AOE Full paper deadline: Jan 28, 2026 AOE Main Track โžœ https://t.co/CYBD7dxFJv Position Papers โžœ
0
0
27
@stefan_fee
Pengfei Liu
3 days
First principle of Context Engineering: Human-Machine Intelligence Gap โ€” Humans naturally "fill in the blanks," machines don't. Context Engineering is fundamentally about entropy reduction, translating high-entropy human intent into machine-understandable signals. Every
@rryssf_
Robert Youssef
4 days
๐Ÿšจ RIP โ€œPrompt Engineering.โ€ The GAIR team just dropped Context Engineering 2.0 โ€” and it completely reframes how we think about humanโ€“AI interaction. Forget prompts. Forget โ€œfew-shot.โ€ Context is the real interface. Hereโ€™s the core idea: โ€œA person is the sum of their
1
5
18
@victorialslocum
Victoria Slocum
2 days
๐—ฆ๐˜๐—ผ๐—ฝ ๐˜„๐—ฟ๐—ถ๐˜๐—ถ๐—ป๐—ด ๐—ฏ๐—ฒ๐˜๐˜๐—ฒ๐—ฟ ๐—ฝ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜๐˜€. Start engineering the system that feeds your LLM the right context at the right time. We've just released our new e-book on ๐—–๐—ผ๐—ป๐˜๐—ฒ๐˜…๐˜ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด going into details on exactly this ๐Ÿ”ฝ Download it for free
16
71
429
@jpineau1
Joelle Pineau
3 days
Cohereโ€™s models and the fabulous @JayAlammar are hard at work, to help us explore all that NeurIPS 2025 has to offer!
@JayAlammar
Jay Alammar
4 days
The Illustrated NeurIPS 2025: A Visual Map of the AI Frontier New blog post! NeurIPS 2025 papers are outโ€”and itโ€™s a lot to take in.ย This visualization lets you explore the entire research landscape interactively, with clusters, summaries, and @cohere LLM-generated explanations
1
11
84
@Dr_Singularity
Dr Singularity
3 days
This seems like a major breakthrough for AI advancement Tencent and Tsinghua introduced CALM (Continuous Autoregressive Language Models), a new approach that replaces next token prediction with continuous vector prediction, allowing the model to think in ideas instead of words.
@rryssf_
Robert Youssef
3 days
Holy shit... this might be the next big paradigm shift in AI. ๐Ÿคฏ Tencent + Tsinghua just dropped a paper called Continuous Autoregressive Language Models (CALM) and it basically kills the โ€œnext-tokenโ€ paradigm every LLM is built on. Instead of predicting one token at a time,
63
239
2K
@aiwithmayank
Mayank Vora
3 days
Prompt Engineering is dead. The GAIR team just dropped Context Engineering 2.0 and it completely reframes how we think about humanโ€“AI interaction. Forget prompts. Forget few-shot. Context is the real interface. Their core idea: โ€œA person is the sum of their contexts.โ€
32
53
202
@iScienceLuvr
Tanishq Mathew Abraham, Ph.D.
3 days
Google DeepMind release: Towards Robust Mathematical Reasoning Introduces IMO-Bench, a suite of advanced reasoning benchmarks that played a crucial role in GDM's IMO-gold journey. Vetted by a panel of IMO medalists and mathematicians. IMO-AnswerBench - a large-scale test on
25
154
977
@Aurimas_Gr
Aurimas Griciลซnas
4 days
๐—”๐—œ ๐—”๐—ด๐—ฒ๐—ป๐˜โ€™๐˜€ ๐— ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† is the most important piece of ๐—–๐—ผ๐—ป๐˜๐—ฒ๐˜…๐˜ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด, this is how we define it ๐Ÿ‘‡ In general, the memory for an agent is something that we provide via context in the prompt passed to LLM that helps the agent to better plan and
44
111
751
@schwarzjn_
Jonathan Richard Schwarz
4 days
This was almost 4 years before the term "Foundational model" was even coined. Early Continual Learning research was genuinely ahead of its time. It was also nice to see an early synergy of nascent ideas in this paper (Bayesian CL + replay through Coresets) ๐Ÿง 
@liyzhen2
yingzhen
4 days
The VCL paper has arguably the first example of modern continual learning for GenAI: VAEs trained on digit/alphabet images 1-by-1 https://t.co/iiMQtOOAt2 Coded by yours truly โ˜บ๏ธ who was (and still is) ๐Ÿฅฐ in generative models. Time to get back to continual learning again?
0
1
6
@rryssf_
Robert Youssef
4 days
๐Ÿšจ RIP โ€œPrompt Engineering.โ€ The GAIR team just dropped Context Engineering 2.0 โ€” and it completely reframes how we think about humanโ€“AI interaction. Forget prompts. Forget โ€œfew-shot.โ€ Context is the real interface. Hereโ€™s the core idea: โ€œA person is the sum of their
106
401
2K
@ChrisLaubAI
Chris Laub
6 days
This broke my brain. A team at Sea AI Lab just discovered that most of the chaos in reinforcement learning training collapse, unstable gradients, inference drift wasnโ€™t caused by the algorithms at all. It was caused by numerical precision. The default BF16 format, used across
19
54
257
@_sungmin_cha
Sungmin Cha
22 days
How can we be sure a generative model (LLMs, Diffusion) has truly unlearned something? What if existing evaluation metrics are misleading us? In our new paper, we introduce FADE, a new metric that assesses genuine unlearning by measuring distributional alignment, moving beyond
1
10
38
@rohanpaul_ai
Rohan Paul
5 days
This thought is converging from many sides. Transformer based LLMs are not going take us to human level AI. That famous Yann LeCun interview. "We are not going to get to human level AI by just scaling up MLMs. This is just not going to happen. There's no way. Okay, absolutely
@rohanpaul_ai
Rohan Paul
6 days
Fei-Fei Li (@drfeifei) on limitations of LLMs. "There's no language out there in nature. You don't go out in nature and there's words written in the sky for you.. There is a 3D world that follows laws of physics." Language is purely generated signal. https://t.co/FOomRpGTad
117
167
1K
@_sungmin_cha
Sungmin Cha
6 days
@prfsanjeevarora As a researcher deeply interested in unlearning, this is a fascinating paper! Your "path-dependence" theory explains why true RE is theoretically impossible & why we observed the "universal failure" in our new work. We proposed FADE, a metric measuring distributional similarity
Tweet card summary image
arxiv.org
Current unlearning metrics for generative models evaluate success based on reference responses or classifier outputs rather than assessing the core objective: whether the unlearned model behaves...
0
1
5
@prfsanjeevarora
Sanjeev Arora
7 days
Very excited with this new finding from our lab: machine unlearning as currently defined (i.e. model should behave as if it had never seen the unlearned data) may be impossible. Main reason: we show that outcome of unlearning --mathematically speaking, gradient ascent --is
@jiatongy_
Jiatong Yu
9 days
Can AI truly forgets? Machine Unlearning (MU) aims to make AI behave as if it has never seen some training data. Our paper: ๐Ÿšจ MU may be impossible as currently conceived ๐Ÿšจ. Paper: https://t.co/hQOHp5S7Rk Homepage: https://t.co/zUyyD1Cq92
11
18
146