Harish Tayyar Madabushi
@harish
Followers
2K
Following
3K
Media
132
Statuses
2K
Lecturer (~Assistant Professor) in Artificial Intelligence. Work on Deep Learning for #NLProc and Deep Contextual Meaning Representations
Bath, England
Joined December 2008
Hey this is me! Our paper: Llama See, Llama Do: A Mechanistic Perspective on Contextual Entrainment and Distraction in LLMs Blog post:
frankniujc.github.io
2
3
22
Iโll be presenting today at 11:00 in hall x5 booth 209 #ACL2025NLP come and letโs talk about how to train with CoTs!
At first I was not sure๐ค, but on second thought, I knew what to do!!!๐ก๐ ๐ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐ฆ๐น on Wednesday 30th at 11:00 #NLProc A ๐งต๐
1
2
10
Excited to present Diverse Chains of Thought at #ACL2025NLP Do you have a dataset with more than one CoT/question? Do you want to train with that? Come to our poster to see how to make the most out of your data! ๐๏ธ Wednesday 30th ๐ฆ11:00 ๐Level 1 1.86 https://t.co/EHTYYSuBqw
underline.io
On-demand video platform giving you access to lectures from conferences worldwide.
At first I was not sure๐ค, but on second thought, I knew what to do!!!๐ก๐ ๐ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐ฆ๐น on Wednesday 30th at 11:00 #NLProc A ๐งต๐
0
1
22
@HaritzPuerto @UKPLab @BathNLP @IGurevych We provide open access to our code, models, data, and results: ๐ฝ๏ธUnderline: https://t.co/fEbrjBXa9l ๐Paper: https://t.co/RNZbxSMPMX ๐ป Code: https://t.co/MlmCFA40Rz ๐ค Models: https://t.co/6bDBxQT84O ๐ Data: https://t.co/XJoB8F1FXB ๐ Website: https://t.co/gdUIWuK0om (9/๐งต)
huggingface.co
0
1
2
@HaritzPuerto @UKPLab @BathNLP @IGurevych We also observed that when we generate 3 CoTs, if the first 2 CoTs are โ and the 3rd is โ
, the model picks the last one! ๐ This shows that DCoT is not an ensemble of CoTs and instead is doing self-correction ๐ 8/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych Why does it work? DCoT attempts to generate subsequent correct CoTs. Maybe the first CoT is wrong โ (and the model doesnโt know it), but by trying to generate a second better CoT, the model may correct the first one โ
๐คฉ 7/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych Generating a second CoT is enough to achieve gains. Note that DCoT@1 remains the same as the vanilla CoT, i.e., training on DCoT is a better way to train an LLM if you have more than one CoT per question. (Both methods were trained with the same CoTs) 6/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych What did we find? Fine-tuning LLMs with DCoT datasets significantly improves performance across all model sizes from 1.3B to 70B parameters. ๐ 5/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych We train CoT and DCoT models with the CoTs. The only difference is that DCoT forces the model to generate them sequentially in a single inference step. With this, we wondered whether LMs can refine their reasoning on the go. 4/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych We created a specialized DCoT dataset, where every question has multiple correct chains of thought. These alternative reasoning paths are all tied to the same answer, encouraging the model to explore diverse solutions simultaneously. ๐คโก๏ธ๐ก 3/๐งต
1
0
1
@HaritzPuerto @UKPLab @BathNLP @IGurevych Traditional CoT methods focus on a single chain of reasoning to arrive at a solution. DCoT, on the other hand, requires models to generate โก๏ธmultiple reasoning paths before producing a final answer, ๐all in a single inference step. 2/๐งต
1
0
1
At first I was not sure๐ค, but on second thought, I knew what to do!!!๐ก๐ ๐ข Diverse Chains of Thought help LLMs refine their Reasoning!! @haritzpuerto will be presenting our work at #ACL2025NLP ๐ฆ๐น on Wednesday 30th at 11:00 #NLProc A ๐งต๐
1
4
16
The trial data has just been released to registered participants. Thereโs still time for your team to join! #emnlp2025 #nlproc
๐ข Call for Participation: TSAR 2025 Shared Task ๐ข Simplify English paragraphs to a specified CEFR level ๐น No training data ๐น Eval: CEFR match, meaning preservation, ref similarity ๐๏ธ Starts July 16 ๐ Info + registration: https://t.co/ydfrHxrBHt
#EMNLP2025 #TextSimplification
0
4
3
๐ขJob Opportunity Research Associate for Reasoning in LLMs, University of Bath, UK (Deadline 05 August 2025) We are looking to hire a highly motivated researcher to work on analysing reasoning in LLMs For moreย information, see:ย https://t.co/2bYI0RglSl
0
11
24
The Cardiff #NLProc Workshop starts on Monday! If you've registered, you should have received a confirmation email (from me). Check your spam folder if not, or feel free to get in touch. Looking forward to seeing many of you in Cardiff!
๐ The Cardiff NLP Workshop kicks off this Monday (14 July)! Check out the full schedule on our website ๐ https://t.co/R3CewFKLP9 Weโve got an exciting lineup: โจ Talks by fantastic speakers ๐ป A tutorial on steering vectors ๐ผ๏ธ Poster session and networking opportunities ๐ An
0
4
11
I am looking for a postdoctoral research associate to work on (LLM-based and neurosymbolic) reasoning for story understanding, as part of the EPSRC-funded ReStoRe project. Details: https://t.co/3yLx5WOviV (deadline 21st July) @Cardiff_NLP @cardiff_krr
0
10
16
Happy to announce our journal paper on tongue twisters, Train and Constrain (TwistList 2.0), has now been officially published in @CompLingJournal! (Thanks to @chenghua_lin and Chen Tang) https://t.co/ecAgSa6vxcโฆ
@sltcdt #nlp #nlproc #nlg
2
4
9
๐จ New global collaboration & dataset paper! UniversalCEFR: Enabling Open Multilingual Research on Language Proficiency Assessment ๐ We introduce UniversalCEFR, an initiative to build a growing, open, multilingual, and multidimensional resource for CEFR-based language
1
9
26
0
0
2