Greg Farquhar Profile
Greg Farquhar

@greg_far

Followers
435
Following
160
Media
5
Statuses
23

Joined October 2017
Don't wanna be here? Send us removal request.
@greg_far
Greg Farquhar
4 years
There’s huge potential in using ‘demonstrations’ from other agents with different goals: to understand which features & dynamics of the environment *might* be important to you; and to borrow from others' behaviours only where they are useful for you.
@filangelos
Angelos Filos
4 years
👽 PsiPhi-learning 👽 (long talk #ICML) https://t.co/TA7gDtEHak shows how an agent can use data from the behavior of other agents with diverse goals: to infer their intentions and fulfill its own! 🧵
1
1
6
@greg_far
Greg Farquhar
4 years
There are a bunch of ideas in this paper, but it all fits together really neatly! Great work from @filangelos and team 👏
0
0
3
@greg_far
Greg Farquhar
5 years
Permanent damage to generalisation from early updates in non-stationary training -- really enjoyed looking into this intriguing problem and trying to solve it for deep RL agents!
@MaxiIgl
Maximilian Igl
5 years
Really excited about our new work: In deep RL, we typically collect new data using a non-stationary policy that gets updated as we learn and improve. We show this can impact the learning dynamics of our deep policy and lead to worse generalization https://t.co/1YTfpzDZOd (1/7)
0
2
15
@greg_far
Greg Farquhar
6 years
This is awesome, but I'm a little scared of how much time I might spend playing it myself...
@_rockt
Tim Rocktäschel
6 years
I am proud to announce the release of the NetHack Learning Environment (NLE)! NetHack is an extremely difficult procedurally-generated grid-world dungeon-crawl game that strikes a great balance between complexity and speed for single-agent reinforcement learning research. 1/
0
0
7
@_rockt
Tim Rocktäschel
6 years
I am proud to announce the release of the NetHack Learning Environment (NLE)! NetHack is an extremely difficult procedurally-generated grid-world dungeon-crawl game that strikes a great balance between complexity and speed for single-agent reinforcement learning research. 1/
14
185
700
@greg_far
Greg Farquhar
6 years
I particularly enjoyed visualising & analysing the learned mixing functions that combine per-agent utilities into joint values!
@_samvelyan
Mikayel Samvelyan
6 years
Happy to share the extended version of our #QMIX paper “Monotonic Value Function Factorisation for Deep Multi-Agent RL” We include further analysis and ablation studies that investigate how monotonic factorisation of joint Q-val helps QMIX outperform VDN https://t.co/AGGADZgumu
0
0
2
@greg_far
Greg Farquhar
6 years
Potential for cool applications in meta-learning, multi-agent learning, etc. If you have ideas or want to chat, let me know or find me at NeurIPS 😀
0
0
7
@greg_far
Greg Farquhar
6 years
A much-improved 🎲Loaded DiCE🎲 objective lets you easily compute low-variance estimators of any-order derivatives for RL. Paper https://t.co/dllhrHuzwD and code https://t.co/NqZsdZy3iT online, nice working with @shimon8282 and @j_foerst! #NeurIPS2019
1
12
61
@polynoamial
Noam Brown
6 years
Tuomas Sandholm and I are doing a Reddit AMA now on the #Pluribus poker AI! https://t.co/qOnCXFSJwe
0
4
25
@greg_far
Greg Farquhar
6 years
AI accelerates by 10x in the hour it takes to repost from r/machinelearning to r/singularityisnear... just how near is it at that rate?? 😱
1
1
13
@greg_far
Greg Farquhar
6 years
Progressively growing the action space creates a great curriculum for learning agents -- check out our paper: https://t.co/YoKe9ZIjhk + code: https://t.co/BdZjplNNEg. Great working with Laura Gustafson @ebetica @shimon8282 Nicolas Usunier @syhw
0
32
130
@_rockt
Tim Rocktäschel
6 years
How can RL agents exploit the compositional, relational and hierarchical structure of the world? A growing number of authors propose learning from natural language. We are excited to share our @IJCAIconf survey of this emerging field! https://t.co/XLHnXMQbVY TL;DR:🤖+📖=📈🎯🏆🥳
2
71
249
@_rockt
Tim Rocktäschel
7 years
I had the pleasure to co-supervise outstanding MSc students jointly with Jakob Foerster (@j_foerst) and Greg Farquhar (@greg_far) at @CompSciOxford this year. Together, we compiled our advice for embarking on short-term machine learning research projects:
3
88
269
@MaxiIgl
Maximilian Igl
7 years
I am very excited to share our ICML paper “Deep Variational Reinforcement Learning (DVRL) for POMDPs”: Our agent learns a model of the environment and acts based on its belief state in this model. w/ @zinmalu @tuananhle7 @frankdonaldwood @shimon8282 https://t.co/XWh5QZ1saU
0
34
122
@shimon8282
Shimon Whiteson
8 years
Our latest paper: how to learn complex joint value functions for teams of agents whose greedy policies can be computed and executed in a decentralised fashion. The trick is a new monotonic value function factorisation. With results on StartCraft 2!
0
32
98
@greg_far
Greg Farquhar
8 years
The camera-ready of our #ICLR2018 paper “TreeQN and ATreeC: Differentiable Tree-Structured Models for Deep Reinforcement Learning” is now online https://t.co/sAUPHT91ho. Code is available at https://t.co/W5iRn8RLD4 @_rockt @MaxiIgl @shimon8282 @whi_rl
0
17
56
@j_foerst
Jakob Foerster
8 years
Excited to share "DiCE: The Infinitely Differentiable Monte Carlo Estimator": https://t.co/LPEy67rCF0 Try this one weird objective for correct any-order gradient estimators in all your stochastic graphs ;) With fantastic Oxford/CMU team: @greg_far @alshedivat @_rockt @shimon8282
3
76
230