
Jason Weston
@jaseweston
Followers
13K
Following
2K
Media
168
Statuses
437
@MetaAI+NYU. NLP from scratch(Pretrain+FT LLM) 2008, MemNets (pre-Transformer) 2015, DrQA(pre-RAG) 2017, BlenderBot(dialog pre-ChatGPT) 2018+,Self-Reward+ more!
NYC
Joined April 2008
RT @danieljwkim: Can we improve Llama 3’s reasoning abilities through post-training only?.Introducing ASTRO, our new framework that teaches….
0
46
0
RT @thao_nguyen26: Web data, the “fossil fuel of AI”, is being exhausted. What’s next?🤔.We propose Recycling the Web to break the data wall….
0
59
0
Reasoning, Attention & Memory Workshop @ COLM.Submission Deadline: June 23, 2025 -- Today!.
🚨Announcing RAM 2 workshop @ COLM25 - call for papers🚨 .- 10 years on, we present the sequel to the classic RAM🐏 (Reasoning, Attention, Memory) workshop that took place in 2015 at the cusp of major change in the area. Now in 2025 we reflect on what's happened and discuss the
3
15
65
RT @chhaviyadav_: Upon graduation, I paused to reflect on what my PhD had truly taught me. Was it just how to write papers, respond to brut….
0
40
0
RT @YifeiZhou02: 📢 New Preprint: Self-Challenging Agent (SCA) 📢. It’s costly to scale agent tasks with reliable verifiers. In SCA, the key….
0
36
0
RT @tesatory: The idea of challenging yourself has a long history, e.g. our Asymmetric Self-Play paper
0
2
0
Interesting work! Also provides additional evidence that our ScPO (Self-Consistency Preference Optimization) direction (Maj vote-based rewards, see fig below) works quite well -- without any labels. @ArchikiPrasad.
🤯 We cracked RLVR with. Random Rewards?!.Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:.- Random rewards: +21%.- Incorrect rewards: +25%.- (FYI) Ground-truth rewards: + 28.8%.How could this even work⁉️ Here's why: 🧵.Blogpost:
0
4
32
RT @kchonyc: oh . what a memory! thanks, @jaseweston et al., for organizing this sequel!. looking back to my slide deck then . i was righ….
0
4
0