oren_sultan Profile Banner
Oren Sultan Profile
Oren Sultan

@oren_sultan

Followers
1K
Following
2K
Media
79
Statuses
621

Research Scientist Intern @Meta, @AIatMeta (FAIR), CS PhD Candidate @HebrewU, @HyadataLab | Past: @Lightricks @TU_Muenchen @UniMelb

Tel Aviv, Israel
Joined August 2021
Don't wanna be here? Send us removal request.
@oren_sultan
Oren Sultan
3 months
Iโ€™m excited to start a new chapter as a PhD Research Scientist Intern at Meta AI, FAIR (Fundamental AI Research) group! Grateful to be part of the CodeGen team in Tel Aviv, working on cutting-edge AI research for code reasoning, understanding and generation ๐Ÿ’ป๐Ÿค–
3
2
106
@naftalibennett
Naftali Bennett ื ืคืชืœื™ ื‘ื ื˜
9 days
ืจืคื™ ื‘ืŸ ืฉื˜ืจื™ืช, ืœืฉืขื‘ืจ ืจืืฉ ื”ืขื™ืจ ื‘ื™ืช ืฉืืŸ, ื”ื•ื ืื‘ื™ื• ืฉืœ ื”ืœื•ื—ื ืกืžืดืจ ืืœืจื•ืื™ ื–ืดืœ ืฉื ืคืœ ื‘ืงืจื‘ ื‘ืžื•ืฆื‘ ื ื—ืœ ืขื•ื–. ืืœืจื•ืื™ ื•ื—ื‘ืจื™ื• ื”ืชืจื™ืขื• ืžืจืืฉ ืฉื—ืžืืก ืžืชื›ื ืŸ ืžืœื—ืžื”, ืืš ืœื ืฉืžืขื• ืœื”ื. ื”ืงืฉื™ื‘ื• ืœื“ื‘ืจื™ื ืžื“ื ืœื™ื‘ื• ืฉืœ ืจืคื™ ื”ื™ืงืจ:
99
196
2K
@moranmiz
Moran Mizrahi
15 days
Looking forward to presenting our TACL paper on enhancing LLM creativity at #EMNLP2025 tomorrow (Wed, Nov 5)! ๐Ÿ“ Room A108 ๐Ÿ• 14:30โ€“16:00 (Linguistic Theories, Cognitive Modeling & Psycholinguistics) Details below โฌ‡๏ธ #NLP #LLMs #Creativity
@moranmiz
Moran Mizrahi
25 days
How can we help LLMs move beyond the obvious toward generating more creative and diverse ideas? In our new TACL paper, we propose a novel approach to enhance LLM creative generation! https://t.co/AFCpQddN6j @ChenShani2 @GabiStanovsky @jurafsky @HyadataLab @stanfordnlp @nlphuji
4
14
60
@uriberger88
Uri Berger
20 days
Heading to #EMNLP2025! ๐ŸŽ‰ Two of our papers will be there โ€” come say hi ๐Ÿ‘‹ ๐Ÿ–ผ๏ธ Image Captioning Evaluation โ€” Nov 5, 17:45 ๐Ÿ“„ https://t.co/TdMVA2iWSD ๐Ÿ•ต๏ธ Deceptive LLM Agents (Mafia Game) โ€” Nov 5, 13:00 ๐Ÿ“„
Tweet card summary image
arxiv.org
LLMs are used predominantly in synchronous communication, where a human user and a model communicate in alternating turns. In contrast, many real-world settings are asynchronous. For example, in...
1
6
26
@guy_yariv
Guy Yariv
26 days
We present DyPE, a framework for ultra high resolution image generation. DyPE adjusts positional embeddings to evolve dynamically with the spectral progression of diffusion. This lets pre-trained DiTs create images with 16M+ pixels without retraining or extra inference cost. ๐Ÿงต๐Ÿ‘‡
9
32
102
@moranmiz
Moran Mizrahi
25 days
How can we help LLMs move beyond the obvious toward generating more creative and diverse ideas? In our new TACL paper, we propose a novel approach to enhance LLM creative generation! https://t.co/AFCpQddN6j @ChenShani2 @GabiStanovsky @jurafsky @HyadataLab @stanfordnlp @nlphuji
6
26
84
@EliahuHorwitz
Eliahu Horwitz
2 months
Excited to share this has now been accepted at #NeurIPS2025 as a position paper (<6% acceptance)!๐ŸŽ‰ We advocate for systematically studying entire model populations via weight-space learning, and argue that this requires charting them in a Model Atlas. @NeurIPSConf #NeurIPS ๐Ÿงต๐Ÿ‘‡
@EliahuHorwitz
Eliahu Horwitz
8 months
๐Ÿšจ New paper alert! ๐Ÿšจ Millions of neural networks now populate public repositories like Hugging Face ๐Ÿค—, but most lack documentation. So, we decided to build an Atlas ๐Ÿ—บ๏ธ Project: https://t.co/1JpsC6dCeg Demo: https://t.co/4Xy7yLdIZY ๐Ÿงต๐Ÿ‘‡๐Ÿป Here's what we found:
0
21
64
@ylecun
Yann LeCun
2 months
Code World Model: producing code by imagining the effect of executing instructions and planning instructions that produce the desired effect.
@syhw
Gabriel Synnaeve
2 months
(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. https://t.co/BJSUCh2vtg
72
177
2K
@adiyossLC
Yossi Adi
2 months
We release Code World Model (CWM)! ๐Ÿ‘ฉโ€๐Ÿ’ป๐ŸŒŽ๐Ÿ“Š A coding LLM designed to advance code generation research through agentic reasoning and world-model-based planning. Super excited about this release and proud of the teamโ€™s work! ๐Ÿ˜ƒ See Gab's post for more info ๐Ÿ‘‡
@syhw
Gabriel Synnaeve
2 months
(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. https://t.co/BJSUCh2vtg
0
11
49
@FelixKreuk
Felix Kreuk
2 months
1/ We released CWM, a 32B dense LLM for coding, agentic use, and, more importantly, to further World-Modeling research. To support this research, we release the pre-training, sft and rl model weights, along with inference code and the tech report. See:
@syhw
Gabriel Synnaeve
2 months
(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. https://t.co/BJSUCh2vtg
1
7
38
@PierreChambon6
Pierre Chambon
2 months
๐Ÿ”ฅ CWM x BigO(Bench) ๐Ÿ”ฅ CWM 32B was just released, and evaluated on BigO(Bench) ! Does "world-modeling-aware" training helps CWM reach higher performance on Code Complexity related tasks ?
2
5
25
@MichaelHassid
Michael Hassid
2 months
Our new Code World Model (CWM) is out! I learned and gained expertise working on the RL part, and I'm super proud of what we built. Check out the thread below for the full details.
@syhw
Gabriel Synnaeve
2 months
(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. https://t.co/BJSUCh2vtg
0
1
15
@AIatMeta
AI at Meta
2 months
New from Meta FAIR: Code World Model (CWM), a 32B-parameter research model designed to explore how world models can transform code generation and reasoning about code. We believe in advancing research in world modeling and are sharing CWM under a research license to help empower
103
225
1K
@alexandr_wang
Alexandr Wang
2 months
new research from Meta FAIR: Code World Model (CWM), a 32B research model we encourage the research community to research this open-weight model! pass@1 evals, for the curious: 65.8 % on SWE-bench Verified 68.6 % on LiveCodeBench 96.6 % on Math-500 76.0 % on AIME 2024 ๐Ÿงต
96
164
1K
@syhw
Gabriel Synnaeve
2 months
(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. https://t.co/BJSUCh2vtg
60
313
2K
@NoySternlicht
Noy Sternlicht๐Ÿ“๐Ÿ—บ๏ธ EMNLP2025
2 months
๐ŸŽ‰ Proud to share that "Debatable Intelligence" has now been accepted to #EMNLP2025 (Main Conference)! https://t.co/zVE73m9lVu Huge thenks to my amazing collaborators @ArielGera2, @RoyBarHaim, @Hoper_Tom, @noamslonim
noy-sternlicht.github.io
We assess the judgment capabilities and behavior of LLMs by analyzing how they rate debate speeches - long texts that argue for or against a controversial topic.
@NoySternlicht
Noy Sternlicht๐Ÿ“๐Ÿ—บ๏ธ EMNLP2025
5 months
๐Ÿ”” New Paper! We propose a challenging new benchmark for LLM judges: Evaluating debate speeches. Are they comparable to humans? Well... itโ€™s debatable. ๐Ÿค” https://t.co/u0sd8SrGjj ๐Ÿ‘‡ Here are our findings:
3
15
54
@EliyaHabba
Eliya Habba @EMNLP ๐Ÿ‡จ๐Ÿ‡ณ
3 months
Proud to share PromptSuite! ๐ŸŒˆ A flexible framework for generating thousands of prompt variations per instance, enabling robust multi-prompt LLM evaluation across diverse tasks. Python API & web UI included. Check it out:
eliyahabba.github.io
A flexible framework for automatic generation of prompt variations for robust LLM evaluation.
@Dahan_Noam
Noam Dahan
3 months
Old news: Single-prompt eval is unreliable๐Ÿคฏ New news: PromptSuite๐ŸŒˆ - an easy way to augment your benchmark with thousands of paraphrases โžก๏ธ robust eval, zero sweat! - Works on any dataset! - Python API + web UI @EliyaHabba, @GiliLior, @GabiStanovsky https://t.co/C4VwIvzJFX
0
2
14
@DavidDinkevich
David Dinkevich
3 months
[1/6] ๐ŸŽฌ New paper: Story2Board We guide diffusion models to generate consistent, expressive storyboards--no training needed. By mixing attention-aligned tokens across panels, we reinforce character identity without hurting layout diversity. ๐ŸŒ https://t.co/aRG81nu5qK
5
11
30
@AsafYehudai
Asaf Yehudai
4 months
๐Ÿšจ Benchmarks tell us which model is better โ€” but not why it fails. For developers, this means tedious, manual error analysis. We're bridging that gap. Meet CLEAR: an open-source tool for actionable error analysis of LLMs. ๐Ÿงต๐Ÿ‘‡
1
14
44
@EliyaHabba
Eliya Habba @EMNLP ๐Ÿ‡จ๐Ÿ‡ณ
4 months
Presenting my poster : ๐Ÿ•Š๏ธ DOVE - A large-scale multi-dimensional predictions dataset towards meaningful LLM evaluation, Monday 18:00 Vienna, #ACL2025 Come chat about LLM evaluation, prompt sensitivity, and our 250M COLLECTION OF MODEL OUTPUTS!
2
11
47