OATML_Oxford Profile Banner
OATML_Oxford Profile
OATML_Oxford

@OATML_Oxford

Followers
5K
Following
192
Media
5
Statuses
207

Oxford Applied and Theoretical Machine Learning Group

Oxford, England
Joined August 2018
Don't wanna be here? Send us removal request.
@ruben_weitzman
Ruben Weitzman
3 months
Huge thanks to the team for this great collaboration between @DeboraMarksLab and @OATML_Oxford with @NotinPascal @PeterM_rchGroth @lood_ml @deboramarks @yaringal. (11/12)
1
1
7
@alxndrdavies
Xander Davies
3 months
Work led by @zhilifeng and @YixuanEvenXu, + @AlexRobey23 and advised by @A_v_i__S and @zicokolter--happy to have contributed to it from @AISecurityInst (along with @_robertkirk) and @OATML_Oxford (along with @yaringal).
0
1
8
@SaFeMMAI
SaFeMM-AI Workshop @ ICCV25
3 months
📢 We are excited to announce Prof. Yarin Gal as a keynote speaker at our #SaFeMMAI workshop at #ICCV2025! Prof. @yaringal is an Associate Professor at the University of Oxford, leads the @OATML_Oxford, and is Director of Research at the UK Government's AI Safety Institute.
Tweet media one
1
5
12
@yaringal
Yarin
6 months
We have a postdoc position available (closing 17th March) to lead work on Generative AI. Please share with anyone you think this might be relevant to! https://t.co/EJONHPXyxO
Tweet media one
0
6
20
@Yoshua_Bengio
Yoshua Bengio
8 months
Very interesting paper about unlearning for AI Safety, a subject that deserves more attention. ⬇️
@FazlBarez
Fazl Barez
8 months
🚨 New Paper Alert: Open Problem in Machine Unlearning for AI Safety 🚨 Can AI truly "forget"? While unlearning promises data removal, controlling emergent capabilities is a inherent challenge. Here's why it matters: 👇 Paper: https://t.co/DXZu17YVeE 1/8
Tweet media one
4
61
324
@StephenLCasper
Cas (Stephen Casper)
7 months
📣 New paper AI gov. frameworks are being designed to rely on rigorous assessments of capabilities & risks. But risk evals are [still] pretty bad – they regularly fail to find overtly harmful behaviors that surface post-deployment. Model tampering attacks can help with this.
Tweet media one
4
44
158
@alxndrdavies
Xander Davies
7 months
Defending against adversarial prompts is hard; defending against fine-tuning API attacks is much harder. In our new @AISecurityInst pre-print, we break alignment and extract harmful info using entirely benign and natural interactions during fine-tuning & inference. 😮 🧵 1/10
Tweet media one
3
23
127
@angusjnic
Angus Nicolson
1 year
📢 New Paper Alert! 📝 "Explaining Explainability: Understanding Concept Activation Vectors" 📄 https://t.co/jRbeY9Zahg What does it mean to represent a concept as a vector? We explore three key properties of concept vectors and how they can affect model interpretations. 🧵(1/5)
Tweet media one
2
33
175
@angusjnic
Angus Nicolson
1 year
📢Competition results! 📢 Very excited to announce TextCAVs, an adaptation of TCAV which can provide concept-based explanations for deep learning models with no data. In the #SaTML2024 competition we found all four secret trojans and came third in the rediscovery benchmark! 🎉
@StephenLCasper
Cas (Stephen Casper)
1 year
📢 Announcing the #SaTML2024 CNN Interpretability Competition results: 🥇 *New record* - Yun et al. - RFLA-Gen2 🥈 Moore et al. - FEUD 🥉 Nicolson - TextCAVs 👏 Tagade and Rumbelow - Prototype Gen Report: https://t.co/FHJmSrd6x2 All were impressive!🧵 on new innovations below.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
4
16
@OATML_Oxford
OATML_Oxford
2 years
Registration for OATML’s Virtual Open Day closes tomorrow. If you are interested in joining for a PhD, we encourage you to register and join us 22 Nov from 4-5:30GMT! Sign-up: https://t.co/1lXOTTMuax Group members attending:
Tweet media one
0
9
24
@OATML_Oxford
OATML_Oxford
2 years
We’re organising a virtual open day! 📅 4-5:30 PM GMT on 22 Nov 2023 Anyone interested in joining @OATML_Oxford 🥣🌾for a PhD is welcome! Come to meet OATMLers, get a feel for our group, and ask any questions you may have! ✏️ Sign Up:
Tweet card summary image
docs.google.com
🌾 We're excited to invite you to our virtual Open Day where you can get to know some OATMLers, learn more about our group, and ask any questions you may have. The event will take place on Wednesday,...
1
13
21
@mplsoxford
MPLSOxford
2 years
New AI tool could help predict viral outbreaks https://t.co/hOLlYmIpWs @OATML_Oxford #AI
Tweet media one
0
1
3
@yaringal
Yarin
2 years
Our awesome Project Manager at our AI/ML research group @OATML_Oxford (based at @CompSciOxford @UniofOxford) will leave us soon :( We're looking for someone to take the role: if you know anyone who's interested, please forward the job advert! closing 21/4 https://t.co/nCSAr06caO
Tweet media one
0
3
12
@alkalait
Freddie Kalaitzis
3 years
I'm very proud that my collaboration w @NASAAmes & the @SETIInstitute is finally published in @NatureAstronomy We show a framework to search for life on another planet in a data-driven fashion, starting from orbit & all the way through to a rover's view https://t.co/t80mZKCZYM
Tweet card summary image
nature.com
Nature Astronomy - A nested orbit-to-ground approach for microbial landscape patterns at different scales, tested in the high Andes, provides a machine learning-based search tool for detecting...
0
1
12
@_lorenzkuhn
Lorenz Kuhn
3 years
How can we measure how uncertain LLMs are about their generations? In our spotlight at #iclr2023 with @yaringal & @sebfar, we introduce "semantic entropy", the entropy over meanings rather than sequences, and show that it reliably measures LM uncertainty🧵 https://t.co/6o3sHDRyqQ
Tweet card summary image
arxiv.org
We introduce a method to measure uncertainty in large language models. For tasks like question answering, it is essential to know when we can trust the natural language outputs of foundation...
5
29
184
@NotinPascal
Pascal Notin
3 years
📢 We are presenting our TranceptEVE preprint at the LMRL workshop today (4-5pm ET, poster # C-48) 📢
Tweet media one
1
8
33
@yaringal
Yarin
3 years
@OATML_Oxford is co-hosting Schmidt AI in Science Fellows to work cutting-edge AI to tackle the big scientific questions of our time. Take a look at some of the research we've done with collaborators in the sciences: https://t.co/Rp0MxR549F
Tweet card summary image
nature.com
Nature - A new computational method, EVE, classifies human genetic variants in disease genes using deep generative models trained solely on evolutionary sequences.
1
1
5