UMass_NLP Profile Banner
UMassNLP Profile
UMassNLP

@UMass_NLP

Followers
1K
Following
191
Media
2
Statuses
94

Natural language processing group at UMass Amherst @umasscs. Led by @thompson_laure @MohitIyyer @brendan642 @andrewmccallum #nlproc

Joined August 2021
Don't wanna be here? Send us removal request.
@kalpeshk2011
Kalpesh Krishna
1 year
@GoogleAI Paper's HuggingFace page:
huggingface.co
0
3
18
@prateeky2806
Prateek Yadav
1 year
Given that I have been closely working with @tuvllms and @kalpeshk2011, I can say that he is extremely well read, hard working and this paper is amazing. People should definitely check out the FLAMe as this is going to be impactful.
@tuvllms
Tu Vu
1 year
🚨 New @GoogleDeepMind paper 🚨 We trained Foundational Large Autorater Models (FLAMe) on extensive human evaluations, achieving the best RewardBench perf. among generative models trained solely on permissive data, surpassing both GPT-4 & 4o. πŸ“°: https://t.co/FIPFiHwXyt 🧡:πŸ‘‡
Tweet media one
Tweet media two
1
6
32
@tuvllms
Tu Vu
1 year
🚨 New @GoogleDeepMind paper 🚨 We trained Foundational Large Autorater Models (FLAMe) on extensive human evaluations, achieving the best RewardBench perf. among generative models trained solely on permissive data, surpassing both GPT-4 & 4o. πŸ“°: https://t.co/FIPFiHwXyt 🧡:πŸ‘‡
Tweet media one
Tweet media two
28
101
565
@kalpeshk2011
Kalpesh Krishna
1 year
Check out our new @GoogleAI paper: we curate a mixture of 5M human judgments to train general-purpose foundational autoraters. Strong LLM-as-judge scores on RewardBench (87.8%), and highest perf among baselines on LLMAggreFact + 6 other benchmarks! πŸ“° https://t.co/oUN4hDeNWx πŸ‘‡
Tweet card summary image
arxiv.org
As large language models (LLMs) advance, it becomes more challenging to reliably evaluate their output due to the high costs of human evaluation. To make progress towards better LLM autoraters, we...
@tuvllms
Tu Vu
1 year
🚨 New @GoogleDeepMind paper 🚨 We trained Foundational Large Autorater Models (FLAMe) on extensive human evaluations, achieving the best RewardBench perf. among generative models trained solely on permissive data, surpassing both GPT-4 & 4o. πŸ“°: https://t.co/FIPFiHwXyt 🧡:πŸ‘‡
Tweet media one
Tweet media two
6
21
124
@MohitIyyer
Mohit Iyyer
1 year
So proud to have hooded my first five PhDs today: @tuvllms, @kalpeshk2011, @simeng_ssun, @mrdrozdov, and Nader Akoury. Now, they're either training LLMs at Google, Nvidia, and Databricks, or staying in academia at Virginia Tech and Cornell. Excited to watch their careers blossom!
Tweet media one
10
9
268
@Anki98765
Ankita Gupta
2 years
Check out ezCoref, our open-source tool for easy coreference annotation across languages/domains. Demo: https://t.co/KIJuYh2aeS Re-annotation study via ezCoref reveals interesting deviations from prior work. πŸ“œ https://t.co/v5tcRck0k9 #CRAC2023 @emnlpmeeting Dec 6, 2:50PM πŸ§΅πŸ‘‡
Tweet media one
1
7
21
@tuvllms
Tu Vu
2 years
πŸ“’ 🌟PhD Openings🌟: I am recruiting PhD students this cycle at Virginia Tech. If you want to dive into: - in-context learning & tool-use LLMs - instruction tuning - parameter-efficient transfer learning - few-shot learning please apply by Dec 15! πŸ‘‰ https://t.co/HiFjcIPSak
5
74
317
@mrdrozdov
Andrew Drozdov
2 years
✨ New Paper ✨ Deep dive on demonstrations to enhance LLM-based passage ranking πŸš€ insights for pointwise ranking using query likelihood πŸš€ https://t.co/KLga6EEx19
Tweet card summary image
huggingface.co
4
26
104
@tuvllms
Tu Vu
2 years
πŸ“’ Want to adapt your outdated LLM to our ever-changing world? 🌏 Check out our code for FreshPrompt at https://t.co/TZcJligS3Q. Colab: https://t.co/ZeKbsjg8n8. πŸ™ We are grateful to @serp_api for their generous sponsorship of 5000 searches for FreshPrompt's users.
Tweet media one
@tuvllms
Tu Vu
2 years
🚨 New @GoogleAI paper: πŸ€– LLMs are game-changers, but can they help us navigate a constantly changing world? πŸ€” As of now, our work shows that LLMs, no matter their size, struggle when it comes to fast-changing knowledge & false premises. πŸ“°: https://t.co/4AkLERtw3q πŸ‘‡
Tweet media one
Tweet media two
0
6
31
@_jasonwei
Jason Wei
2 years
Nice paper by Tu Vu on factuality in LLMs: https://t.co/JNe4eU4pHy, enjoyed contributing in a minor role to it while I was at Google. The main takeaway for me is that most factuality benchmarks for LLMs don't really take into account the fact that many types of knowledge
@tuvllms
Tu Vu
2 years
🚨 New @GoogleAI paper: πŸ€– LLMs are game-changers, but can they help us navigate a constantly changing world? πŸ€” As of now, our work shows that LLMs, no matter their size, struggle when it comes to fast-changing knowledge & false premises. πŸ“°: https://t.co/4AkLERtw3q πŸ‘‡
Tweet media one
Tweet media two
3
15
115
@YapeiChang
Yapei Chang
2 years
Can LLMs summarize books exceeding their context windows? We design an evaluation protocol for collecting fine-grained human judgments on LLM-generated summaries & propose BooookScore, a reference-free automatic metric for narrative coherence. https://t.co/xNXToBzt9r 🧡below:
Tweet card summary image
arxiv.org
Summarizing book-length documents (>100K tokens) that exceed the context window size of large language models (LLMs) requires first breaking the input document into smaller chunks and then...
2
48
224
@quocleix
Quoc Le
2 years
A weakness of LLMs is that they don’t know recent events well. This is nice work from Tu developing a benchmark (FreshQA) to measure factuality of recent events, and a simple method to improve search integration for better performance on the benchmark.
@tuvllms
Tu Vu
2 years
🚨 New @GoogleAI paper: πŸ€– LLMs are game-changers, but can they help us navigate a constantly changing world? πŸ€” As of now, our work shows that LLMs, no matter their size, struggle when it comes to fast-changing knowledge & false premises. πŸ“°: https://t.co/4AkLERtw3q πŸ‘‡
Tweet media one
Tweet media two
1
7
47
@tuvllms
Tu Vu
2 years
🚨 New @GoogleAI paper: πŸ€– LLMs are game-changers, but can they help us navigate a constantly changing world? πŸ€” As of now, our work shows that LLMs, no matter their size, struggle when it comes to fast-changing knowledge & false premises. πŸ“°: https://t.co/4AkLERtw3q πŸ‘‡
Tweet media one
Tweet media two
5
86
381
@MohitIyyer
Mohit Iyyer
2 years
Evaluating the factuality of LLMs is tricky: what if they answer a question correctly but also generate a bunch of unrelated made-up stuff? We eval LLM answers to our new FreshQA dataset in both a "strict" (no made up stuff) and "relaxed" setting, see the paper for more!
@tuvllms
Tu Vu
2 years
🚨 New @GoogleAI paper: πŸ€– LLMs are game-changers, but can they help us navigate a constantly changing world? πŸ€” As of now, our work shows that LLMs, no matter their size, struggle when it comes to fast-changing knowledge & false premises. πŸ“°: https://t.co/4AkLERtw3q πŸ‘‡
Tweet media one
Tweet media two
0
7
49
@MSheshera
Sheshera Mysore
2 years
I’m at #sigir2023 and presenting our work on interactively controllable personalisation! Come listen in room 101 between 11-12.30!
@MSheshera
Sheshera Mysore
2 years
Friends, I am happy that I share our #SIGIR2023 paper on learning transparent user profiles for text recommendation! πŸ“œ https://t.co/aEzfSg5CHZ With our model, LACE: Users can examine + interact with their profiles and interactively improve their recommendations. 🧡🧡🧡
Tweet media one
0
3
26
@brendan642
brendan o'connor
2 years
Reminder - for the terrific interdisciplinary Text as Data conference, abstract submissions coming up - due Aug 4! https://t.co/BjegNt4qpv It's a great, small, non-archival conference to discuss emerging work with folks across social sciences, humanities, and computer science.
tada2023.org
0
40
67
@UMass_NLP
UMassNLP
2 years
In Prize-winning Paper, UMass Amherst Computer Scientists Release Guidelines for Evaluating AI-Generated Text : UMass Amherst
2
7
45
@MohitIyyer
Mohit Iyyer
2 years
Huge congrats @tuvuumass, who just became my first graduated PhD student!! He'll be starting his own group soon @VT_CS, so prospective PhD applicants interested in topics like multitask/multimodal transfer learning, or param-efficient LLM adaptation: def apply to work with him!
@tuvllms
Tu Vu
2 years
I successfully defended my Ph.D. thesis. A special thank you to the members of my thesis committee: my wonderful advisor @MohitIyyer, @MajiSubhransu, @HamedZamani, @lmthang, and @colinraffel for their insightful feedback and advice on my research and career plans.
0
5
91
@tuvllms
Tu Vu
2 years
Moving forward, I will be splitting my time as a research scientist at @GoogleAI and an assistant professor @VT_CS. I will also be recruiting Ph.D. students starting in Fall 2024 to work on effective and efficient transfer learning in the era of LLMs, please come join me!
4
7
55
@tuvllms
Tu Vu
2 years
I would also like to thank all of my labmates @UMass_NLP and friends at @UMassAmherst, my mentors and collaborators at @GoogleAI and @MSFTResearch, and my family and friends all over the world who gave me support and encouragement throughout my Ph.D. journey.
1
1
22