yihong_thu Profile Banner
Yihong Chen Profile
Yihong Chen

@yihong_thu

Followers
213
Following
199
Media
1
Statuses
47

The constant happiness is curiosity.

London, UK
Joined May 2015
Don't wanna be here? Send us removal request.
@yihong_thu
Yihong Chen
1 year
RT @LiuChunan: Presented our project "AsEP: Benchmarking Deep Learning Methods for Antibody-specific Epitope Prediction" at #ISMB2024 in Mo….
0
2
0
@yihong_thu
Yihong Chen
1 year
RT @karpathy: @maurosicard This information was never released but I'd expect it was a lot more. In terms of multipliers let's say 3X from….
0
14
0
@grok
Grok
20 hours
Generate videos in just a few seconds. Try Grok Imagine, free for a limited time.
415
173
2K
@yihong_thu
Yihong Chen
1 year
RT @kchonyc: we all want to and need to be prepared to train our own large-scale language models from scratch. why?. 1. transparency or….
0
101
0
@yihong_thu
Yihong Chen
1 year
RT @xinchiqiu: This is a great collaboration with 2 amazing collaborators from Cambridge, UCL and FAIR Meta.@williamfshen and @yihong_thu ;….
0
2
0
@yihong_thu
Yihong Chen
1 year
the link to active forgetting paper is broken in the above text. if you are curious, here it is
0
0
0
@yihong_thu
Yihong Chen
1 year
Active forgetting 2307.01163 ( finds its new frontier with @surajk610 et al's cool work on structural in-context learning for LMs! Scaling makes LMs powerful but forgetting gives them plasticity -- accepting and adapting to novel XYZ quickly. #AI.
@surajk610
Suraj Anand
1 year
How robust are in-context algorithms? In new work with @michael_lepori, @jack_merullo, and @brown_nlp, we explore why in-context learning disappears over training and fails on rare and unseen tokens. We also introduce a training intervention that fixes these failures.
Tweet media one
1
0
9
@yihong_thu
Yihong Chen
1 year
RT @niclane7: This week at @iclr_conf catch up with Wanru (@Renee42581826), PhD student @CaMLSys and @Cambridge_CL, who has led some import….
0
2
0
@yihong_thu
Yihong Chen
1 year
RT @robertarail: I’ll be at #ICLR2024 on Saturday for the LLM Agents Workshop Panel 🚀. Some of my collaborators will also be there through….
0
10
0
@yihong_thu
Yihong Chen
1 year
RT @flwrlabs: 🌼 Flower AI Summit 2024: The world's largest federated learning conference is about to start 🏃🏼‍♀️🏃🏽. 📺 Watch live now: https….
0
3
0
@yihong_thu
Yihong Chen
1 year
RT @Kseniase_: How to teach a language model a new language without retraining?. The key: periodic forgetting. A recent research paper dem….
0
86
0
@yihong_thu
Yihong Chen
1 year
Quanta covers our work on regular embedding resetting for language plasticity. Worth noting that language plasticity is not the only type of plasticity current language models need. Wish to see more insights into AI plasticity from @nikishin_evg @oh_that_hat @clarelyle.
@QuantaMagazine
Quanta Magazine
1 year
To learn more flexibly, a new machine learning model selectively forgets what it already knows. @settostun reports:
1
7
23
@yihong_thu
Yihong Chen
1 year
I am very sure @riedelcastro beats AI in this sense!.
@riedelcastro
Sebastian Riedel (@[email protected])
1 year
I often feel that the main reason I made it through school was my quite extraordinary forgetfulness. I really had to rederive everything all the time. @yihong_thu showed how this trait can also help language models in her NeurIPS paper Now on Quanta!.
0
0
10
@yihong_thu
Yihong Chen
2 years
RT @ylecun: @shimon8282 If you really want a large collection of ugly slide decks, look here:.
0
27
0
@yihong_thu
Yihong Chen
2 years
RT @zhu_zhaocheng: 📜 Check out our latest blog post where we reflect on the past year's research in tool use and reasoning with Large Langu….
Tweet card summary image
towardsdatascience.com
It's the beginning of 2024 and ChatGPT just celebrated its one-year birthday. One year is a super long time for the community of large...
0
11
0
@yihong_thu
Yihong Chen
2 years
and almost forgot the paper link 😅
0
2
8
@yihong_thu
Yihong Chen
2 years
work done with a couple of amazing folks: @cheeesio, @robertarail, @davlanade, Pontus Stenetorp, @riedelcastro and @artetxem.
1
0
5
@yihong_thu
Yihong Chen
2 years
If you're deciding which #NeurIPS23 poster to check out tomorrow, don't forget our forgetting paper! Visit poster #328 Thursday morning to dive into the world of active forgetting. Discover how it enhances language models with greater language plasticity. See you there!
Tweet media one
Tweet media two
Tweet media three
2
14
48
@yihong_thu
Yihong Chen
2 years
RT @oanacamb: 🚨💫We are delighted to have @shishirpatil_ at our @uclcs NLP Meetup *Monday 1st Nov 6:30pm GMT* . The event will be *hybrid*….
0
7
0
@yihong_thu
Yihong Chen
2 years
RT @lena_voita: [1/7] A new paper (and a blog post!) with @javifer_96 and @christofernal!. Neurons in LLMs: Dead, N-gram, Positional. Blog:….
0
65
0