Yao Lu Profile
Yao Lu

@yaolu_nlp

Followers
235
Following
89
Media
2
Statuses
31

PhD Student @ucl_nlp, former member of @UWaterloo @Mila_Quebec and @AmiiThinks

London
Joined February 2017
Don't wanna be here? Send us removal request.
@yaolu_nlp
Yao Lu
3 years
Excited to receive an ACL outstanding paper award, with @max_nlp @latticecut @riedelcastro @ucl_nlp ! TL;DR If prompting is not working, change the order, the performance may jump from random-guess to SOTA. How to find fantastically ordered prompts? Here➡️ https://t.co/R6NktEebwJ
5
32
203
@soheeyang_
Sohee Yang
1 year
🚨 New Paper 🚨 Can LLMs perform latent multi-hop reasoning without exploiting shortcuts? We find the answer is yes – they can recall and compose facts not seen together in training or guessing the answer, but success greatly depends on the type of the bridge entity (80%+ for
7
48
202
@yaolu_nlp
Yao Lu
1 year
🎉🎉🏆
0
0
0
@lintool
Jimmy Lin
1 year
They say a picture is worth a thousand words... but work led by @ralph_tang finds words worth a thousand pictures! https://t.co/L8wU715wRd
1
5
22
@LoubnaBenAllal1
Loubna Ben Allal
1 year
🍷 FineWeb technical report is out and so is 📚 FineWeb-Edu, a 1.3 trillion tokens dataset that outperforms all other open web datasets, with remarkable improvements on educational benchmarks such as MMLU, ARC, and OpenBookQA. Technical report: https://t.co/lfOZYYJKxq Dataset:
10
61
266
@yaolu_nlp
Yao Lu
1 year
@ralph_tang @riedelcastro (3/3) 📈 Compared to the standard templates widely used in LLM benchmarks, our random prompt baseline yielded an average relative improvement of over 10%! Take a look at our work https://t.co/MHac6nMg2J, unlocking the potential of language model optimisation space!
1
0
10
@yaolu_nlp
Yao Lu
1 year
@ralph_tang @riedelcastro (2/3) we're outperforming human experts! This establishes a robust baseline for prompt optimisation. Even for basic text classification tasks, Large Language Models (such as LLaMA, Mistral, ChatGPT) can experience significant improvement using #python random prompt optimisation
1
1
2
@yaolu_nlp
Yao Lu
1 year
(1/3) Excited to introduce our #NAACL2024 paper "Strings from the Library of Babel: Random Sampling as a Strong Baseline for Prompt Optimisation". We've unlocked the power of #python random library for prompt optimisation✨With a simple trick - random.sample(vocabulary) as prompt
4
13
58
@WecoAI
Weco AI
2 years
We're excited to announce AIDE has become the first human-level AI agent for data science! AIDE outperforms half of human data scientists on a wide range of Kaggle competitions, surpassing conventional AutoML, LangChain agents, and ChatGPT with human assistance. 🏆
7
54
226
@yaolu_nlp
Yao Lu
2 years
Congrats on the launch! LFG @YuxiangJWu @zhengyaojiang
@WecoAI
Weco AI
2 years
Today we’re announcing Weco AI and our first product, AIDE: your AI agent for Machine Learning. Simply describe your task in natural language, and AIDE will search the design space to deliver source code and a report for you. Join the waitlist now at https://t.co/VWhju2ohZY (1/3)
0
0
2
@gneubig
Graham Neubig
2 years
Often prompt engineering focuses on the *content* of the prompt, but in reality *formatting* of the prompt can have an equal or larger effect, especially for less powerful models. This is a great deep dive into this phenomenon by @melaniesclar et al.
@melaniesclar
Melanie Sclar
2 years
Did you know that depending on the format used in few-shot prompting, you may get accuracies ranging 4%-88% for a given task w/LLaMA-2-70B 5-shot? or 47%-85% w/GPT3.5?🤯 We explore this variance in FormatSpread, or: How I learned to start worrying about prompt formatting. 1/n
1
10
113
@a_stadt
Alex Warstadt
2 years
LLMs are now trained >1000x as much language data as a child, so what happens when you train a "BabyLM" on just 100M words? The proceedings of the BabyLM Challenge are now out along with our summary of key findings from 31 submissions: https://t.co/zli0jzA1XP Some highlights 🧵
13
188
1K
@charliebholtz
Charlie Holtz
2 years
David Attenborough is now narrating my life Here's a GPT-4-vision + @elevenlabsio python script so you can star in your own Planet Earth:
711
5K
26K
@oanacamb
Oana-Maria Camburu
2 years
🚨💫We are delighted to have @shishirpatil_ at our @uclcs NLP Meetup *Monday 1st Nov 6:30pm GMT* The event will be *hybrid* Due to room capacity, there are *two links* to sign up depending on whether you attend in person or online Details in: https://t.co/pcnWPmDt8F
0
7
11
@_yuxiangwu
Yuxiang (Jimmy) Wu
3 years
Introducing ChatArena 🏟 - a Python library of multi-agent language game environments that facilitates communication and collaboration between multiple large language models (LLMs)! 🌐🤖 Check out our GitHub repo: https://t.co/xisQrxeQr7 #ChatArena #NLP #AI #LLM 1/8 🧵
13
139
594
@ucl_nlp
UCL Natural Language Processing
5 years
We would like to wish you all a restful winter break, and best wishes for the new year 🙂🎄
0
3
30
@ml_perception
Mike Lewis
5 years
Happy to share MARGE, our new work on rethinking pre-training: given a document, we first retrieve related documents, and then paraphrase these to reconstruct the original. MARGE works well for generation and classification in many languages, sometimes without supervision. (1/6)
8
89
398
@marinkazitnik
Marinka Zitnik
6 years
We will present a tutorial on Machine Learning for Drug Development at #IJCAI2020! Materials to follow on our website: https://t.co/w0JZJhef8B @IJCAIconf #drugs #networks #AI
1
25
65
@rsalakhu
Russ Salakhutdinov
6 years
#ICLR2020 paper on Differentiable Reasoning over a Virtual Knowledge Base: Efficient, end-to-end differentiable framework for doing complex multi-hop QA over a large text corpus. https://t.co/P60JiWv3EC w/t Dhingra, Zaheer, Balachandran, @gneubig , @professorwcohen
0
51
191
@evolvingstuff
evolvingstuff
6 years
Transformers as Soft Reasoners over Language "we explore whether transformers can similarly learn to reason (or emulate reasoning), but using rules expressed in language, thus bypassing a formal representation." https://t.co/5uib4bRPud Datasets and demo: https://t.co/wKxgdCpHiE
6
48
179