harish Profile Banner
Harish Tayyar Madabushi Profile
Harish Tayyar Madabushi

@harish

Followers
2K
Following
3K
Media
132
Statuses
2K

Lecturer (~Assistant Professor) in Artificial Intelligence. Work on Deep Learning for #NLProc and Deep Contextual Meaning Representations

Bath, England
Joined December 2008
Don't wanna be here? Send us removal request.
@frankniujc
Jingcheng (Frank) Niu
4 months
Hey this is me! Our paper: Llama See, Llama Do: A Mechanistic Perspective on Contextual Entrainment and Distraction in LLMs Blog post:
frankniujc.github.io
2
3
22
@HaritzPuerto
Haritz Puerto
4 months
Iโ€™ll be presenting today at 11:00 in hall x5 booth 209 #ACL2025NLP come and letโ€™s talk about how to train with CoTs!
@harish
Harish Tayyar Madabushi
4 months
At first I was not sure๐Ÿค”, but on second thought, I knew what to do!!!๐Ÿ’ก๐Ÿ˜ƒ ๐Ÿ“ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐Ÿ‡ฆ๐Ÿ‡น on Wednesday 30th at 11:00 #NLProc A ๐Ÿงต๐Ÿ‘‡
1
2
10
@HaritzPuerto
Haritz Puerto
4 months
Excited to present Diverse Chains of Thought at #ACL2025NLP Do you have a dataset with more than one CoT/question? Do you want to train with that? Come to our poster to see how to make the most out of your data! ๐Ÿ—“๏ธ Wednesday 30th ๐Ÿ•ฆ11:00 ๐Ÿ“Level 1 1.86 https://t.co/EHTYYSuBqw
Tweet card summary image
underline.io
On-demand video platform giving you access to lectures from conferences worldwide.
@harish
Harish Tayyar Madabushi
4 months
At first I was not sure๐Ÿค”, but on second thought, I knew what to do!!!๐Ÿ’ก๐Ÿ˜ƒ ๐Ÿ“ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐Ÿ‡ฆ๐Ÿ‡น on Wednesday 30th at 11:00 #NLProc A ๐Ÿงต๐Ÿ‘‡
0
1
22
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych We provide open access to our code, models, data, and results: ๐Ÿ“ฝ๏ธUnderline: https://t.co/fEbrjBXa9l ๐Ÿ“„Paper: https://t.co/RNZbxSMPMX ๐Ÿ’ป Code: https://t.co/MlmCFA40Rz ๐Ÿค— Models: https://t.co/6bDBxQT84O ๐Ÿ“‚ Data: https://t.co/XJoB8F1FXB ๐ŸŒ Website: https://t.co/gdUIWuK0om (9/๐Ÿงต)
Tweet card summary image
huggingface.co
0
1
2
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych We also observed that when we generate 3 CoTs, if the first 2 CoTs are โŒ and the 3rd is โœ…, the model picks the last one! ๐ŸŽ‰ This shows that DCoT is not an ensemble of CoTs and instead is doing self-correction ๐ŸŽŠ 8/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych Why does it work? DCoT attempts to generate subsequent correct CoTs. Maybe the first CoT is wrong โŒ (and the model doesnโ€™t know it), but by trying to generate a second better CoT, the model may correct the first one โœ…๐Ÿคฉ 7/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych Generating a second CoT is enough to achieve gains. Note that DCoT@1 remains the same as the vanilla CoT, i.e., training on DCoT is a better way to train an LLM if you have more than one CoT per question. (Both methods were trained with the same CoTs) 6/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych What did we find? Fine-tuning LLMs with DCoT datasets significantly improves performance across all model sizes from 1.3B to 70B parameters. ๐ŸŽ‰ 5/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych We train CoT and DCoT models with the CoTs. The only difference is that DCoT forces the model to generate them sequentially in a single inference step. With this, we wondered whether LMs can refine their reasoning on the go. 4/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych We created a specialized DCoT dataset, where every question has multiple correct chains of thought. These alternative reasoning paths are all tied to the same answer, encouraging the model to explore diverse solutions simultaneously. ๐Ÿค”โžก๏ธ๐Ÿ’ก 3/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
@HaritzPuerto @UKPLab @BathNLP @IGurevych Traditional CoT methods focus on a single chain of reasoning to arrive at a solution. DCoT, on the other hand, requires models to generate โžก๏ธmultiple reasoning paths before producing a final answer, ๐Ÿ”„all in a single inference step. 2/๐Ÿงต
1
0
1
@harish
Harish Tayyar Madabushi
4 months
At first I was not sure๐Ÿค”, but on second thought, I knew what to do!!!๐Ÿ’ก๐Ÿ˜ƒ ๐Ÿ“ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐Ÿ‡ฆ๐Ÿ‡น on Wednesday 30th at 11:00 #NLProc A ๐Ÿงต๐Ÿ‘‡
1
4
16
@feralvam
Fernando Alva-Manchego
4 months
The trial data has just been released to registered participants. Thereโ€™s still time for your team to join! #emnlp2025 #nlproc
@feralvam
Fernando Alva-Manchego
4 months
๐Ÿ“ข Call for Participation: TSAR 2025 Shared Task ๐Ÿ“ข Simplify English paragraphs to a specified CEFR level ๐Ÿ”น No training data ๐Ÿ”น Eval: CEFR match, meaning preservation, ref similarity ๐Ÿ—“๏ธ Starts July 16 ๐Ÿ”— Info + registration: https://t.co/ydfrHxrBHt #EMNLP2025 #TextSimplification
0
4
3
@harish
Harish Tayyar Madabushi
4 months
๐Ÿ“ขJob Opportunity Research Associate for Reasoning in LLMs, University of Bath, UK (Deadline 05 August 2025) We are looking to hire a highly motivated researcher to work on analysing reasoning in LLMs For moreย information, see:ย  https://t.co/2bYI0RglSl
0
11
24
@nedjmaou
Nedjma Ousidhoum ู†ุฌู…ุฉ ุฃูˆุณูŠุฏู‡ู…
4 months
The Cardiff #NLProc Workshop starts on Monday! If you've registered, you should have received a confirmation email (from me). Check your spam folder if not, or feel free to get in touch. Looking forward to seeing many of you in Cardiff!
@Cardiff_NLP
Cardiff NLP
4 months
๐ŸŽ‰ The Cardiff NLP Workshop kicks off this Monday (14 July)! Check out the full schedule on our website ๐Ÿ‘‰ https://t.co/R3CewFKLP9 Weโ€™ve got an exciting lineup: โœจ Talks by fantastic speakers ๐Ÿ’ป A tutorial on steering vectors ๐Ÿ–ผ๏ธ Poster session and networking opportunities ๐Ÿ”„ An
0
4
11
@StevenSchockae2
Steven Schockaert
5 months
I am looking for a postdoctoral research associate to work on (LLM-based and neurosymbolic) reasoning for story understanding, as part of the EPSRC-funded ReStoRe project. Details: https://t.co/3yLx5WOviV (deadline 21st July) @Cardiff_NLP @cardiff_krr
0
10
16
@tylerl404
๐•ฟ๐–ž๐–‘๐–Š๐–— ๐•ท๐–”๐–†๐–๐–’๐–†๐–“
5 months
Happy to announce our journal paper on tongue twisters, Train and Constrain (TwistList 2.0), has now been officially published in @CompLingJournal! (Thanks to @chenghua_lin and Chen Tang) https://t.co/ecAgSa6vxcโ€ฆ @sltcdt #nlp #nlproc #nlg
2
4
9
@josephimperial_
Joseph Imperial
5 months
๐Ÿšจ New global collaboration & dataset paper! UniversalCEFR: Enabling Open Multilingual Research on Language Proficiency Assessment ๐ŸŒ We introduce UniversalCEFR, an initiative to build a growing, open, multilingual, and multidimensional resource for CEFR-based language
1
9
26
@harish
Harish Tayyar Madabushi
6 months
0
0
2
@harish
Harish Tayyar Madabushi
6 months
Starting in 15 minutes! Looking forward to this talk by @DrGeofreyHinton #NLProc
1
1
6