conorfhayes Profile Banner
Conor Hayes Profile
Conor Hayes

@conorfhayes

Followers
158
Following
417
Media
0
Statuses
73

Research scientist @ Cognizant AI Lab

Oakland, CA
Joined September 2017
Don't wanna be here? Send us removal request.
@conorfhayes
Conor Hayes
18 days
🚨🚨 New 📜 from Cognizant AI Lab! 🚨🚨 We show that Evolution Strategies (ES) can scale to fine-tune LLMs with billions of parameters, beating RL on robustness, sample efficiency, and tolerance to long-horizon tasks without gradient computation. See @yule_gan's amazing 🧵below:
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
1
0
4
@yule_gan
Yulu Gan
18 hours
Check out this nice tutorial video ( https://t.co/n63MxIhaso) from @yacinelearning I also did a live chat with him this morning — check out the recording ( https://t.co/aQBOPCIBVb) where I answered some questions from Yacine and the audience about our work :)
@yacinelearning
Yacine Mahdid
3 days
alright we're live in about 3 min to figure out how the guys manage to make evolutionary strategies works for finetuning LLMs tune in!
0
4
23
@conorfhayes
Conor Hayes
1 day
Join to ask questions about our paper with @yule_gan 🔥🔥
@yacinelearning
Yacine Mahdid
3 days
ladies and gentleman this thursday at 10:00 AM EST we are going to run a Q&A with @yule_gan one of the author of that nice LLM finetuning paper with evolution strategies tune in to ask him any dumb questions you might have on ES, RL, tickling LLMs, or what's next.
0
0
5
@yacinelearning
Yacine Mahdid
3 days
ladies and gentleman this thursday at 10:00 AM EST we are going to run a Q&A with @yule_gan one of the author of that nice LLM finetuning paper with evolution strategies tune in to ask him any dumb questions you might have on ES, RL, tickling LLMs, or what's next.
@yacinelearning
Yacine Mahdid
5 days
this sunday we are figuring out how folks scaled evolutionary optimization methods to LLMs kinda cool that old tricks are used again in modern times to great effects
5
2
78
@yacinelearning
Yacine Mahdid
3 days
we are live over here folks for the live paper review on evolution strategies at scale:
1
1
11
@yacinelearning
Yacine Mahdid
5 days
this sunday we are figuring out how folks scaled evolutionary optimization methods to LLMs kinda cool that old tricks are used again in modern times to great effects
19
22
457
@josancamon19
Joan Cabezas
14 days
🧵 As AI labs race to scale RL, one question matters: when should you stop pre-training and start RL? We trained 5 Qwen models (0.6B→14B) with RL on GSM8K and found something wild: Small models see EMERGENCE-LIKE jumps. Large models see diminishing returns. The scaling law?
38
120
758
@apsarathchandar
Sarath Chandar
17 days
If you are attending @COLM_conf and looking to hire a research scientist, I highly recommend you talk to my postdoc, Mathieu Reymond, who is in the job market and at the conference! Mathieu is an expert in mult-objective RL, multi-agent RL, RL for scientific discovery, and RL for
2
12
37
@martinbowling
Martin Bowling
18 days
If ES really beats PPO/GRPO on reasoning this could be super compelling. Nice work @yule_gan & team.
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
0
1
4
@hardmaru
hardmaru
17 days
Evolution Strategies can be applied at scale to fine-tune LLMs, and outperforms PPO and GRPO in many model settings! Fantastic paper “Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning” by @yule_gan, Risto Miikkulainen and team. https://t.co/CEyX6Z5ulG
Tweet card summary image
arxiv.org
Fine-tuning pre-trained large language models (LLMs) for down-stream tasks is a critical step in the AI deployment pipeline. Reinforcement learning (RL) is arguably the most prominent fine-tuning...
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
10
36
279
@aomaru_21490
Jiaxin Ge @ ICCV25
18 days
Super cool work that shows finetuning with Evolution Strategies (ES) in the parameter space outperforms GRPO! Check out
Tweet card summary image
github.com
This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning" - VsonicV/es-fine-tuning-paper
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
1
3
14
@far__el
Far El
18 days
damn never thought my joke last year (noisekit, https://t.co/46fhRupnd3) could be made into a viable training algorithm 😂
Tweet card summary image
github.com
Contribute to pharaouk/noisekit development by creating an account on GitHub.
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
1
1
16
@jeffclune
Jeff Clune
18 days
These are some of my favorite papers I have been a part of, because of the deep insight we learned while making them (driven by @joelbot3000 's genius). Glad to see them inspire cool new work!
@yule_gan
Yulu Gan
18 days
Our work stands on the shoulders of giants: @jeffclune and @kenneth0stanley demonstrated the potential of ES in several very insightful papers, including https://t.co/z5iTMEPc2u and https://t.co/nBCOz3VpQF. Earlier, @SchmidhuberAI proposed Natural Evolution Strategies (NES)
1
4
26
@Jarrattp
Paul Jarratt
18 days
What comes after Reinforcement Learning? Cognizant AI Lab scaled Evolution Strategies (ES) to fine-tune LLMs with billions of parameters — no gradients, less instability, and more efficiency. #finetuningllm #reinforcementlearning A new path forward begins here. 🔗 Blog | Paper
0
1
1
@yule_gan
Yulu Gan
18 days
Our work stands on the shoulders of giants: @jeffclune and @kenneth0stanley demonstrated the potential of ES in several very insightful papers, including https://t.co/z5iTMEPc2u and https://t.co/nBCOz3VpQF. Earlier, @SchmidhuberAI proposed Natural Evolution Strategies (NES)
2
8
121
@kenneth0stanley
Kenneth Stanley
18 days
Nice to see an exploration of the potential for ES (evolution strategies) in LLM fine tuning! Many potential advantages are discussed in this thread from @yule_gan .
@yule_gan
Yulu Gan
18 days
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for
4
16
147
@yule_gan
Yulu Gan
18 days
Thanks for sharing! @rohanpaul_ai We hope our paper provides insights into a new direction for LLM fine-tuning
@rohanpaul_ai
Rohan Paul
18 days
The paper shows that evolution strategies can fine tune full LLMs at scale and often beat reinforcement learning on reasoning. The key finding is that parameter space search with only outcome scores can outperform token level RL across models and tasks. It tweaks whole models,
2
4
19
@robleclerc
Rob Leclerc
18 days
despite the bitter lesson, i’m still convinced that evolutionary strategies—particularly with an elegant genotype-to-phenotype map that directs the development of a neural architecture—will play a major role in future advances in ai.
@rohanpaul_ai
Rohan Paul
18 days
The paper shows that evolution strategies can fine tune full LLMs at scale and often beat reinforcement learning on reasoning. The key finding is that parameter space search with only outcome scores can outperform token level RL across models and tasks. It tweaks whole models,
0
1
2
@rohanpaul_ai
Rohan Paul
18 days
The paper shows that evolution strategies can fine tune full LLMs at scale and often beat reinforcement learning on reasoning. The key finding is that parameter space search with only outcome scores can outperform token level RL across models and tasks. It tweaks whole models,
8
40
237