
Joshua Clymer
@joshua_clymer
Followers
2K
Following
792
Media
44
Statuses
429
Turtle hatchling trying to make it to the ocean. I work at Redwood Research.
Joined April 2022
I'm confused why global AI coordination is so strongly associated with a concentration of power. Global coordination => no single leader => diffusion of power?.
Wait, I finally listened to the Peter Thiel Antichrist interview, and the Antichrist part is literally just the same as my dynamism essay. ?. Peter Thiel me. đ€.Totalitarianism is not a great solution to AI risks, actually
2
0
14
RT @inferencemag: Inference is hosting some of the worldâs leading experts for a debate on the possibility and potential consequences of auâŠ.
0
23
0
How do we know if AI systems are safe from misuse?. The short answer: have a red team try to misuse them, and measure the effort required. But the devil is in the detailsâwhich I dive into in a safety case written with @_robertkirk and others.
New paper! With @joshua_clymer, Jonah Weinbaum and others, weâve written a safety case for safeguards against misuse. We lay out how developers can connect safeguard evaluation results to real-world decisions about how to deploy models. đ§”
5
2
22
RT @AliciaP59828402: Guys I'll read the book but pleaseee - you still have time to change the cover art! đ©
0
5
0
this is bad news. the more nations have lots of AI chips the more difficult coordinating a multi-year slowdown will be.
The US is making the UAE and Saudi Arabia into great AI powers of their own (alongside the US and China) for little apparent benefit other than a few people getting very rich.
2
4
41
I'm glad for more serious investigation of existential risk from think tanks, but I think ASI will be much better at identifying paths to human extinction than these authors. I wish they did not make sweeping claims like "Extinction threats posed by AI are immensely challenging".
A new RAND report on AI x-risk is shockingly bad; I don't see how it got past their internal peer review. There are many issues, but the main critical flaw is the conflation of "It seems hard to me" with "It will be hard for a superintelligent AI". Other issuse:.- Not grappling
0
1
12
RT @NeelNanda5: New post: I'm all for investment in interpretability but IMO this overstates its importance vs other safety methods. I disaâŠ.
0
30
0