MMitchell
@mmitchell_ai
Followers
81K
Following
57K
Media
752
Statuses
22K
Interdisciplinary researcher focused on shaping AI towards long-term positive goals. ML & Ethics. Similar content in the Skies (this bird has flown).
Joined June 2016
My partner: "So is Google Search just basically unusable now? All I can find is garbage."
My feed is such a dynamic mix of (1) people in tech SUPER EXCITED about advancing AI more & more (2) people *not* in tech begging tech people to stop, with very clear details as to why. It's amazing I can be in both worlds; I wish (1) listened (actually listened) to (2) more. đ«€
1
1
9
My feed is such a dynamic mix of (1) people in tech SUPER EXCITED about advancing AI more & more (2) people *not* in tech begging tech people to stop, with very clear details as to why. It's amazing I can be in both worlds; I wish (1) listened (actually listened) to (2) more. đ«€
1
0
6
đ€đ Hot off the presses, just published from @wef! "AI Agents in Action: Foundations for Evaluation and Governance" Honored to have been a part of this effort led by @CedricLarsen and @JasonDePerro with many brilliant people at the forefront of AI. https://t.co/NolyAkE6t1
weforum.org
AI Agents in Action: Foundations for Evaluation and Governance 2025 clarifies how organizations can align AI agent adoption with proportionate safeguards.
2
6
24
"Data are no longer things to be accounted for...but rather inputs to...models". Many in LLM-ML don't care about the problems they are actually building models of: "the nature of languages...how we work w/ language...specific contexts [of use]." https://t.co/cve4r1NIza
medium.com
This is a question I posted on BlueSky on Friday 11/21/25, inspired by a talk I recently attended about evaluation of âAIâ systems. I thinkâŠ
0
0
3
2- Is something everyone knows. This place (X) isn't for a diverse and inclusive discussion of values. There's not much more I can say on this, given what this platform is. /end
1
1
3
You can see why this shorthand is a problem: If someone's values amounts to "kill people" (or whatever), the shorthand would allow them to say "this is ethical". 4/
1
0
3
I've observed 2 things about social-media-level AI ethics discussions: 1-When people say "this is ethical", it's shorthand for having some set of values--prioritizing some, ignoring others--and what they're showing is aligned with those. 3/
1
0
0
That's one of the points behind the "trolley problem". There isn't a right answer. All answers are right in some way, and wrong in some way. 2/
2
1
1
Ethics is about understanding things through different value lenses. Justice, Kindness, &c. There isn't a universal right answer--that's the point of ethics. There are different right answers for different people dependent on their personal (& organizational) proclivities. 1/
2
0
6
As FYI: If someone tells you something theyâre doing is âethicalâ, they donât understand how ethics works. đ§”
6
5
31
If you get AI Slopped this holiday season, would love to see pics! Of this form: AI generated image vs Actual Image
NEW: AI ârecipe slopâ is overrunning search and social. Food creators say Googleâs AI Overviews and glossy fake food pics are drowning out tested recipesâcollapsing traffic and setting home cooks up for disaster, especially this Thanksgiving. Gift link:
1
1
13
Our new paper Economies of Open Intelligence is out and covered by @Melissahei in the @FinancialTimes. It offers the clearest picture yet of how global power is shifting inside the open AI ecosystem, and what it means for the open-source community.
4
16
50
This one is pretty nasty - it tricks Antigravity into stealing AWS credentials from a .env file (working around .gitignore restrictions using cat) and then leaks them to a webhooks debugging site that's included in the Antigravity browser agent's default allow-list
Top of HackerNews today: our article on Google Antigravity exfiltrating .env variables via indirect prompt injection -- even when explicitly prohibited by user settings!
49
338
2K
@weiss_hadas youâre in luck, we have entered âletâs circle back to this in the new year!â season
0
2
390
đ€ Was recently at a trustworthy AI agent event with Rida Fayyaz , and she had a brilliant idea: Add deterministic layers where possible. My take is that these can serve as a kind of 'gate' on agent behavior and access.
2
1
14
New Pew report confirms whatâs seemed obvious to me for awhile: X is not a place one goes to interact w/women âNever in the history of modern social media has one gender so decisively abandoned a platform⊠X male-female imbalance is less extreme only than late-2010s Redditâ 1/2
1
7
22
Pro tip: If the person providing a different interpretation is an expert in a relevant field and not speaking for Moneyed Interests, theyâre an âindependent scholarâânot just a âskepticâ.
0
2
27
One of the changes that OpenAI has made to make ChatGPT safer is a "take a break" nudge. There's something quite interesting about the design here. Which thing does this message make you want to click?
The company essentially turned a dial that made ChatGPT more appealing and made people use it more, but that sent some of them into delusional spirals. OpenAI has since made the chatbot safer, but that came with a tradeoff: less usage.
34
45
371
LLMs are forcing me to change my writing style and i am not happy about it
16
11
299