Steve Omohundro
@steveom
Followers
2K
Following
658
Media
37
Statuses
722
Beneficial AI Research, 2024 Future of Life Award For Pioneering Scholarship in Computer Ethics and AI Safety
Palo Alto, California
Joined June 2008
My essay "Regulating AGI: From Liability to Provable Contracts" https://t.co/R7vIrohJEO was just posted as part of the "AGI Social Contract" project: https://t.co/mZAQ1oj5ft
agisocialcontract.org
A consortium of experts designing the necessary strategies to guide governments through the upcoming AI economic transformation.
1
4
14
AI theorem proving advancing rapidly! Critical to AGI Safety!
Happy to introduce Kimina-Prover-72B ! Reaching 92.2% on miniF2F using Test time RL. It can solve IMO problems using more than 500 lines of Lean 4 code ! Check our blog post here: https://t.co/QbrmoyYL9i And play with our demo ! https://t.co/u0Wj0Id4vZ
0
1
8
Very important work highlighting the current state of AI Cyberattacks, the likely most critical next AI Risk.
1/ 🔥 AI agents are reaching a breakthrough moment in cybersecurity. In our latest work: 🔓 CyberGym: AI agents discovered 15 zero-days in major open-source projects 💰 BountyBench: AI agents solved real-world bug bounty tasks worth tens of thousands of dollars 🤖
0
0
6
Extremely important work! With all the excitement around AI code generation, let's get the ball rolling on verified code generation!
1/🧵Introducing VERINA: a high-quality benchmark for verifiable code generation. As LLMs are increasingly used to generate software, we need more than just working code--We need formal guarantees of correctness. VERINA offers a rigorous and modular framework for evaluating LLMs
1
4
8
🧵 "What should the role of governments be after the upcoming AI economic transformation?" We're launching an new anthology on a new "AGI Social Contract" - in collab with 12+ experts to explore new strategies & policy interventions for an AI transition. In collab with:
12
41
191
A group of leading AI scholars is backing a private governance model for AI policy solutions — and they're rallying around SB 813. 🧵
2
6
26
Great work! You've brought theorem proving to the thinking LLM revolution! Thank you for making the prover and the autoformalizer freely available. I believe these will be essential to effective AI Safety.
We believe formal math is the future. 🔥Introducing Kimina-Prover Preview, a Numina & @Kimi_Moonshot collaboration, the first large formal reasoning model for Lean 4, achieving 80.78% miniF2F. https://t.co/fNX7orQYeZ
0
3
9
Shocking! This year's batch of Y Combinator startups is all in on "Vibe Coding": https://t.co/HHzgs17xLB At 10:00 they say "one quarter of the founders said that more than 95% of their codebase was AI-generated!"
0
1
3
random thoughts/predictions on where vibe coding might go: - most code will be written (generated?) by the time rich. Thus, most code will be written by kids/students rather than software engineers. This is the same trend as video, photos, and other social media - we are in the
152
130
1K
Amazing example of "Vibe Coding" an entire flight simulation game using thousands of prompts and no coding:
Wow. I cannot believe it. Just asked Claude to make the dogfight ultra realist! ✅ hit impacts ✅ smoke when damaged ✅ explosion on death ✅ free-fall with smoke It feels so good to fly! + awesome plane and controls, 100% in Cursor with 0 code edition from me. LOOK AT THIS!
1
0
1
I just gave a short talk arguing that extending "Vibe Coding" to "Vibe Proving" and "Vibe Specification" will power formal methods for AI Safety: "It's a New Day for Formal Methods!"
drive.google.com
0
5
14
I agree that R1 shows that we are unlikely to achieve AI Safety by mandating constraints on the big lab AIs. But "Acceleration is the only way forward" is suicide. My opposite take is: "It's time to get serious about building truly safe and secure infrastructure so that humanity
Whether you like it or not, the future of AI will not be canned genies controlled by a "safety panel". The future of AI is democratization. Every internet rando will run not just o1, but o8, o9 on their toaster laptop. It's the tide of history that we should surf on, not swim
14
18
75
In November, I was honored to participate in the Mind First Foundation's event "AI Safety Salon with Steve Omohundro": https://t.co/9VC6kfoTS1 They just posted the two videos of the event: https://t.co/nDXvB7oZJp
https://t.co/6VOGhZkLGj In addition to Q&A and a panel discussion
0
0
5
AI safety people were right. Again. Instrumental convergence is proven yet again. Empirically. So far we’ve seen: ✅ Self-preservation: You can’t achieve your goals if you’re turned off (see Apollo's recent paper. Link in comments) ✅Resource acquisition: You can’t achieve
18
19
182
🏆 ⭐ We're thrilled to announce the 2024 Future of Life Award winners! This year, we honor three groundbreaking experts who laid the foundations for ethics and safety in computing and AI. Learn more about the invaluable work of Batya Friedman, James Moor, and Steve Omohundro:
5
18
74
PSA for Boston-area folks!! On Nov 22nd there will be a special AI safety salon at Microsoft NERD w/ Steve Omohundro (@steveom), AI safety pioneer. Organized by Alex Hoekstra & @PrestonWEstep from Mind First Foundation (@mind_first). RSVP here-> https://t.co/jLjOb5X2Yu
1
1
6
We need more "tool AI" as in Elon's space triumph today, and less AGI world domination hype as in Dario Amodei's recent "entente" essay. My essay below argues that "scaling quickly" won't lead to Dario's "eternal 1991" – but perhaps to 1984 until the end, with a non-human Big
22
32
169
Debates over AI Policy like CA SB-1047 highlight fragmentation in the AI community. How can we develop AI policies that help foster innovation while mitigating risks? We propose a path for science- and evidence-based AI policy: https://t.co/9G9mlZ6JHi
8
66
251