ManasiSharma_ Profile Banner
Manasi Sharma Profile
Manasi Sharma

@ManasiSharma_

Followers
344
Following
114
Media
13
Statuses
44

research engineer @scale_AI, working on reasoning for frontier models, agents, rl | prev @stanford, @StanfordAILab, @mitll, @Columbia

Stanford, CA
Joined July 2020
Don't wanna be here? Send us removal request.
@ManasiSharma_
Manasi Sharma
2 years
Happy CS Commencement Day! .S/o to Prof. @chrmanning for making CS 224N (Natural Language Processing) one of the most fun and rewarding courses that I have ever TA'd for😊.@stanfordnlp @StanfordEng @Stanford @StanfordAILab
Tweet media one
2
5
61
@ManasiSharma_
Manasi Sharma
1 month
going to ICML next week! let me know if you're interested in chatting about RL, reasoning, agents (esp. browser agents) and data + evals :).
3
2
98
@ManasiSharma_
Manasi Sharma
2 months
new work on rlvr + natural language guidance to enhance a model's ability to reason in its own words.
@SeanHendryx
Sean Hendryx
2 months
What will the learning environments of the future look like that train artificial super intelligence? In recent work at @scale_AI , we show that training systems that combine verifiable rewards with multi-agent interaction accelerate learning.
Tweet media one
Tweet media two
1
1
6
@ManasiSharma_
Manasi Sharma
2 months
first work since joining Scale is out!.
@iScienceLuvr
Tanishq Mathew Abraham, Ph.D.
2 months
Adaptive Guidance Accelerates Reinforcement Learning of Reasoning Models. "Guide adaptively incorporates hints into the model’s context on problems for which all rollouts were initially incorrect and adjusts the importance sampling ratio for the "off-policy" trajectories in order
Tweet media one
2
0
4
@ManasiSharma_
Manasi Sharma
2 months
or apply directly:......
Tweet card summary image
scale.com
0
0
1
@ManasiSharma_
Manasi Sharma
2 months
the Reasoning & Agents team at Scale AI is hiring researchers!🩵we publish research on new forms of supervision & training algorithms for reasoning models, agentic environments, & collaborate with top frontier labs to improve their models. feel free to DM me if interested!
Tweet media one
3
0
2
@ManasiSharma_
Manasi Sharma
5 months
excited to share that i have joined the frontier data research team at @scale_AI!🎉 i will be working on research in the domains of LLM reasoning, agents and rl.
0
0
6
@ManasiSharma_
Manasi Sharma
6 months
RT @askalphaxiv: DeepSeek-R1 now offers an open source model with the ability to explain its intermediate reasoning. But how do explanation….
0
2
0
@ManasiSharma_
Manasi Sharma
7 months
Amazing initiative by @rajpalleti314 and team!.
@askalphaxiv
alphaXiv
7 months
Goodreads for arXiv papers💡. What if instead of arbitrary algorithms and tweets, arXiv papers were curated by your research community?. Introducing communities on alphaXiv: bridging papers, discussions, and people in one space.
1
2
6
@ManasiSharma_
Manasi Sharma
8 months
and @jefrankle for the talk and sick databricks merch!
Tweet media one
0
0
2
@ManasiSharma_
Manasi Sharma
8 months
s/o @shaneguML for a very fun chat on multilinguality (and for taking the best pictures at the conference)
Tweet media one
1
0
2
@ManasiSharma_
Manasi Sharma
8 months
and thanks @polynoamial for answering some of my questions on test-time compute and reasoning
Tweet media one
1
0
1
@ManasiSharma_
Manasi Sharma
8 months
thank you @drfeifei for sharing more about @theworldlabs (and an amazing live demo)
Tweet media one
1
0
1
@ManasiSharma_
Manasi Sharma
8 months
Had a great time at NeurIPS in Vancouver this year and met some amazing people in person!.
1
0
3
@ManasiSharma_
Manasi Sharma
8 months
i’m going to be attending NeurIPS again this year! am super excited to meet new people and chat / learn more about: reasoning, post-training, alignment, agentic systems, code-gen (& please share any cool party invites!).
1
0
11
@ManasiSharma_
Manasi Sharma
10 months
Been a while since I taught 224n, thank you for the kind words :).
@Thor_9111
Sam
10 months
Day 20 of NLP: .- finished cs224n lecture 19 .- it was purely python tutorial .- did a recap on numpy on:. > broadcasting . > matrix multiplication . > hardamard nproduct . @ManasiSharma_ is good instructor.
Tweet media one
Tweet media two
0
0
6
@ManasiSharma_
Manasi Sharma
1 year
Open to any questions or comments!. Authors: Manasi Sharma, Ho Chit Siu, Rohan Paleja, Jaime D. Peña. 8/8.
0
0
1
@ManasiSharma_
Manasi Sharma
1 year
Overall, we find that humans tend to trust differing model explanations equitably (and highly) in isolation, although post-hoc ones are trusted more when compared directly. This suggests important ramifications for human trust in plausibly misinformative model responses. 7/8.
1
0
1
@ManasiSharma_
Manasi Sharma
1 year
We do not find a statistically significant tradeoff between explanation types and model performance, indicating that choosing explanation types that maximize user trust may not negatively affect model performance during text generation. 6/8
Tweet media one
1
0
2
@ManasiSharma_
Manasi Sharma
1 year
However, these gains disappear when users are shown responses independently, suggesting that humans trust all model responses fairly highly (~4/5 on a Likert scale), including deceptive ones, equitably when they are shown in isolation. 5/8
Tweet media one
1
0
2
@ManasiSharma_
Manasi Sharma
1 year
We find that any explanation from the model about its reasoning significantly increases (p ≤ 1.4×10-7) self-reported user trust when the user has the chance to compare responses. Trust is also higher when the justification follows the answer (post-hoc) in the retrieved case. 4/8
Tweet media one
1
0
2