LisaAlazraki Profile Banner
Lisa Alazraki Profile
Lisa Alazraki

@LisaAlazraki

Followers
1K
Following
2K
Media
10
Statuses
94

PhD student @ImperialCollege. Research Scientist Intern @AIatMeta prev. @Cohere, @GoogleAI. Interested in generalisable learning and reasoning. She/her

London, UK
Joined January 2016
Don't wanna be here? Send us removal request.
@LisaAlazraki
Lisa Alazraki
15 days
Just arrived at #EMNLP2025 in Suzhou. Looking forward to meeting with everyone! Will be giving an oral presentation of our paper No Need For Explanations: LLMs Can Implicitly Learn from Mistakes In Context this Friday 7th November at 11.30 am in Hall A108 ๐ŸŽค
1
8
84
@ICComputing
Imperial Computing
12 days
Weโ€™re looking to hire four Assistant/Associate Professors to our Department. Our priority areas are: ๐Ÿ”ต Programming Languages ๐Ÿ”ต Systems ๐Ÿ”ต Security ๐Ÿ”ต Software Engineering ๐Ÿ”ต Computer Architecture ๐Ÿ”ต Theoretical Computer Science Deadline: 15 Dec 2025 https://t.co/VwBxJpJQpe
Tweet card summary image
imperial.ac.uk
Please note that job descriptions are not exhaustive, and you may be asked to take on additional duties that align with the key responsibilities ment...
0
1
6
@MavorParker
Augustine Mavor-Parker
13 days
At Vmax, we are automating the construction of RL environments and the post-training of agents. We are hiring members of technical staff and research fellows. Come join us in SF! (link to apply in comments).
5
30
69
@Cohere_Labs
Cohere Labs
19 days
Cohere Labs x EMNLP 2025: "No Need for Explanations: LLMs can implicitly learn from mistakes in-context" This work demonstrates that eliminating explicit corrective rationales from incorrect answers improves Large Language Models' performance in math reasoning tasks,
0
3
12
@imperial_nlp
Imperial NLP
20 days
Our @imperial_nlp community is ready for #EMNLP2025 โœˆ๏ธ Check out the papers weโ€™ll be presenting next week in Suzhou and come say hi!
0
6
25
@mrl_workshop
Multilingual Representation Workshop @ EMNLP 2025
20 days
Introducing Global PIQA, a new multilingual benchmark for 100+ languages. This benchmark is the outcome of this yearโ€™s MRL shared task, in collaboration with 300+ researchers from 65 countries. This dataset evaluates physical commonsense reasoning in culturally relevant contexts.
2
55
106
@NishantBalepur
Nishant Balepur
21 days
๐Ÿšจ New Paper! ๐Ÿšจ We want ~helpful~ LLMs, and RLHF-ing them with user preferences and reward models will get us there, right? WRONG! ๐Ÿ™…โŒโ›”๏ธ Our #EMNLP2025 paper finds a major helpfulness-preferences gap: user/LLM judgments + agent simulations can totally miss what helps users
7
20
69
@louvishh
lovish
1 month
๐Ÿšจ New Paper: The Art of Scaling Reinforcement Learning Compute for LLMs ๐Ÿšจ We burnt a lot of GPU-hours to provide the community with the first open, large-scale systematic study on RL scaling for LLMs. https://t.co/49REQZ4R6G
@Devvrit_Khatri
Devvrit
1 month
Wish to build scaling laws for RL but not sure how to scale? Or what scales? Or would RL even scale predictably? We introduce: The Art of Scaling Reinforcement Learning Compute for LLMs
2
17
76
@max_nlp
Max Bartolo
1 month
Thanks for having me! Fantastic to see such innovative research happening at @imperialcollege and inspiring to meet so many brilliant students shaping the future of AI research!
@borruell
Borja G. Leon
1 month
Sold out @ic_arl seminar with @max_nlp!
0
3
25
@lorenz_wlf
Lorenz Wolf
1 month
Post-training methods like RLHF improve LLM quality but often collapse diversity. Check out DQO, a training objective using DPPs that directly optimizes for semantic diversity and quality.
@YileiChen49
Yilei Chen
1 month
๐Ÿš€Excited to introduce ๐ƒ๐๐Ž (๐ƒ๐ข๐ฏ๐ž๐ซ๐ฌ๐ข๐ญ๐ฒ ๐๐ฎ๐š๐ฅ๐ข๐ญ๐ฒ ๐Ž๐ฉ๐ญ๐ข๐ฆ๐ข๐ณ๐š๐ญ๐ข๐จ๐ง), a principled method for post-training Large Language Models to generate diverse high-quality responses ๐Ÿ“– https://t.co/xJsvsVz6i1
0
1
6
@Preethi__S_
Preethi Seshadri
1 month
Arjun and Vagrant will be presenting our work at Session 3 (11AM ๐Ÿ•š poster session) at @COLM_conf. If you're interested in LLM evaluation practices, probability vs. generation-based evals, harms and misgendering, please go say hi ๐Ÿ‘‹! Link to paper ๐Ÿ“œ:
Tweet card summary image
arxiv.org
Numerous methods have been proposed to measure LLM misgendering, including probability-based evaluations (e.g., automatically with templatic sentences) and generation-based evaluations (e.g., with...
0
1
8
@liyzhen2
yingzhen
2 months
Available by Dec 2025: Oliver Ratmann and I are looking for 1 PhD student on AI for pathogen deep-sequence analytics. Scholarship includes UK home fee + stipend. Contact my work email if interested. RT ๐Ÿ™ https://t.co/TGg6a1NL52
Tweet card summary image
docs.google.com
AI for pathogen deep-sequence analytics An AI4Health CDT PhD project (student to be recruited by Dec 2025) Scholarship: Home fee + stipend Supervisors (50:50 split) Yingzhen Li (Department of...
0
8
17
@irombie
Irem Ergรผn
2 months
Delighted to announce that our work, ShiQ: Bringing back Bellman to LLMs has been accepted to NeurIPS 2025 ๐Ÿฅณ๐Ÿฅณ๐ŸŽ‰๐ŸŽ‰๐ŸŽŠ๐ŸŽŠ Details of the paper in the thread below ๐Ÿ‘‡๐Ÿ‘‡
@irombie
Irem Ergรผn
6 months
I'm excited to share our new pre-print ShiQ: Bringing back Bellman to LLMs! https://t.co/yWMT6M0nuT In this work, we propose a new, Q-learning inspired RL algorithm for finetuning LLMs ๐ŸŽ‰ (1/n)
3
3
41
@DulhanJay
Dulhan Jayalath
2 months
๐ŸšจNew Meta Superintelligence Labs Paper๐Ÿšจ What do we do when we donโ€™t have reference answers for RL? What if annotations are too expensive or unknown? Compute as Teacher (CaT๐Ÿˆ) turns inference compute into a post-training supervision signal. CaT improves up to 30% even on
13
85
542
@jacobyhsi88
Jacob Si
2 months
VUDโ€™s gonna be at #NeurIPS2025 ๐ŸŽ‰๐Ÿฅณ Special thanks to my labmates that made this collaboration especially enjoyable!
@jacobyhsi88
Jacob Si
2 months
Wanna understand the sources of uncertainty in LLMs when performing in-context learning ๐Ÿค”? ๐Ÿš€ We introduce a variational uncertainty decomposition framework for in-context learning without explicitly sampling from the latent parameter posterior. ๐Ÿ“„ Paper:
0
5
32
@LisaAlazraki
Lisa Alazraki
2 months
โœจ Accepted as a Spotlight at #NeurIPS2025! Huge thanks to my coauthors and everyone who supported us. Check out the details below ๐Ÿ‘‡
@LisaAlazraki
Lisa Alazraki
6 months
Thrilled to share our new preprint on Reinforcement Learning for Reverse Engineering (RLRE) ๐Ÿš€ We demonstrate that human preferences can be reverse engineered effectively by pipelining LLMs to optimise upstream preambles via reinforcement learning ๐Ÿงตโฌ‡๏ธ
12
34
273
@LisaAlazraki
Lisa Alazraki
2 months
This will be an oral! ๐ŸŽค See you at #EMNLP25
@LisaAlazraki
Lisa Alazraki
3 months
This is accepted to EMNLP Main! Looking forward to presenting it in Suzhou ๐ŸŽ‰
1
4
28
@matthieu_meeus
Matthieu Meeus
2 months
I was recently asked during an interview why LLMs were not deterministic even when temperature is 0. This very carefully answers it! My interviewerโ€™s answer was different though, that there likely is non determinism in MOE routing for load balancing
@thinkymachines
Thinking Machines
2 months
Today Thinking Machines Lab is launching our research blog, Connectionism. Our first blog post is โ€œDefeating Nondeterminism in LLM Inferenceโ€ We believe that science is better when shared. Connectionism will cover topics as varied as our research is: from kernel numerics to
1
1
26
@DulhanJay
Dulhan Jayalath
2 months
ChatGPT doesn't understand why I keep beating it at rock paper scissors๐Ÿ˜…
1
1
9
@jacobyhsi88
Jacob Si
2 months
Wanna understand the sources of uncertainty in LLMs when performing in-context learning ๐Ÿค”? ๐Ÿš€ We introduce a variational uncertainty decomposition framework for in-context learning without explicitly sampling from the latent parameter posterior. ๐Ÿ“„ Paper:
@liyzhen2
yingzhen
2 months
We show how to make LLM in-context learning approximately Bayesian & decompose uncertainty IMO this is proper approximate inference ๐Ÿฅฐ applied to LLMs Led by awesome students @shavindra_j @jacobyhsi88 Filippo & Wenlong ๐Ÿ‘ Example๐Ÿ‘‡by prompting, bandits & NLP examples in paper
2
9
24