geoffreyirving Profile Banner
Geoffrey Irving Profile
Geoffrey Irving

@geoffreyirving

Followers
10K
Following
12K
Media
180
Statuses
4K

Chief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.

London
Joined September 2009
Don't wanna be here? Send us removal request.
@geoffreyirving
Geoffrey Irving
27 days
I am very excited that AISI is announcing over £15M in funding for AI alignment and control, in partnership with other governments, industry, VCs, and philanthropists! . Here is a 🧵 about why it is important to bring more independent ideas and expertise into this space.
@AISecurityInst
AI Security Institute
27 days
📢Introducing the Alignment Project: A new fund for research on urgent challenges in AI alignment and control, backed by over £15 million. ▶️ Up to £1 million per project. ▶️ Compute access, venture capital investment, and expert support . Learn more and apply ⬇️.
9
27
165
@geoffreyirving
Geoffrey Irving
3 days
From near the end of Sleepwalkers, by Christopher Clark, as World War I starts.
Tweet media one
0
0
2
@grok
Grok
6 days
What do you want to know?.
546
336
2K
@geoffreyirving
Geoffrey Irving
6 days
The correct mathematical definition is the one that makes the most intermediate lemmas happen to be true, along the way to the result you care about.
3
1
17
@geoffreyirving
Geoffrey Irving
14 days
RT @alxndrdavies: I’ve been wanting to see this result for a while! We find evidence that narrow filtering of pre-training data can degrade….
0
5
0
@geoffreyirving
Geoffrey Irving
14 days
RT @StephenLCasper: 🧵 New paper from @AISecurityInst x @AiEleuther that I led with Kyle O’Brien:. Open-weight LLM safety is both important….
0
39
0
@geoffreyirving
Geoffrey Irving
16 days
RT @michael_nielsen: I'm quite bothered by the last post. It's written in social media advice speak :-(. The underlying point is, I think,….
0
1
0
@geoffreyirving
Geoffrey Irving
18 days
RT @AISecurityInst: The more capable models become, the harder it is to safely evaluate them - but it’s essential that AI agents are tested….
0
13
0
@geoffreyirving
Geoffrey Irving
19 days
RT @alxndrdavies: We at @AISecurityInst worked with @OpenAI to test GPT-5's safeguards. We identified multiple jailbreaks, including a univ….
0
24
0
@geoffreyirving
Geoffrey Irving
19 days
This is the key point about infinities: the world has a lot of big numbers, and it is often way easier to do mathematics if we approximate those big numbers with infinities of one form or another.
@Anthony_Bonato
Anthony Bonato
19 days
To be fair, I don't know if infinities exist in nature. But modern mathematics shows infinities are vg at approximating finite systems.
1
1
8
@geoffreyirving
Geoffrey Irving
20 days
I often see the implicit assumption that if we automate AI safety research, it’s enough to do so much of it that we generate a right answer. But this is not correct: we also need to know that all the other, cheaper, wrong answers are wrong.
1
0
17
@geoffreyirving
Geoffrey Irving
21 days
RT @danielmurfet: Neural networks are grown, not programmed. What does that growth process look like? Like this!. This is a small language….
0
140
0
@geoffreyirving
Geoffrey Irving
24 days
RT @activewarp: get in loser, we’re ranking our favorite fp4 representable values
Tweet media one
0
11
0
@geoffreyirving
Geoffrey Irving
24 days
Thank you, Yoshua! Very excited to get to work together in this new capacity!.
@Yoshua_Bengio
Yoshua Bengio
26 days
Pleased to see this new Alignment Project, where I serve as an expert advisor, launched by the UK's @AISecurityInst and supported by the Canadian AI Safety Institute and many others. I encourage my fellow researchers to apply for funding, compute and support from int’l experts.
0
0
4
@geoffreyirving
Geoffrey Irving
25 days
RT @matt_levine: This newsletter has been an embarrassment internally, but its failure isn’t a big concern.
Tweet card summary image
bloomberg.com
Not legal advice. Also Builder.ai, Boring Co., Harvard and AI.
0
7
0
@geoffreyirving
Geoffrey Irving
26 days
RT @juddrosenblatt: Excellent point from @geoffreyirving :. cognitive science, theory, and clever ML experiments are "massively, systematic….
0
3
0
@geoffreyirving
Geoffrey Irving
26 days
RT @AmmannNora: Very excited to see this come out, and to be able to support! . Beyond the funding itself, the RfP itself is a valuable res….
0
5
0
@geoffreyirving
Geoffrey Irving
27 days
A final note: I am confident that alignment is solvable, but this does not imply that we'll find the solution in time, nor that technical alignment alone is sufficient for safety. We should try! If we don't succeed, something else that follows us will succeed instead.
1
0
13
@geoffreyirving
Geoffrey Irving
27 days
Therefore, a core goal of the Alignment Project is bringing in researchers with relevant expertise who aren't yet working on alignment, both via funds and by helping to map the research space. Our research agenda aims to help here, and we would love to talk to interested people!.
1
0
8
@geoffreyirving
Geoffrey Irving
27 days
There is a lot of risk in trying only one approach to simulating the future! There are only so many SOTA LLM-focused safety research areas. They might succeed, or they might all fail in correlated ways. Spreading our bets helps!.
1
1
10