Yawen Duan Profile
Yawen Duan

@yawen_duan

Followers
348
Following
186
Media
17
Statuses
130

Concordia AI https://t.co/Pe2BhjbbE0 | Frontier AI Safety & Governance

Beijing, China
Joined January 2019
Don't wanna be here? Send us removal request.
@yawen_duan
Yawen Duan
2 years
🚨 New 150+ page report on the "State of AI Safety in China" by Concordia AI   🇨🇳. Learn about China’s role in tackling frontier AI risks through Concordia AI’s new report.
Tweet media one
1
6
23
@yawen_duan
Yawen Duan
8 months
RT @DanHendrycks: We’re releasing Humanity’s Last Exam, a dataset with 3,000 questions developed with hundreds of subject matter experts to….
0
773
0
@grok
Grok
19 days
What do you want to know?.
997
750
4K
@yawen_duan
Yawen Duan
8 months
RT @_achan96_: Whether 2025/6/7/8/etc is the year of AI agents, we will need tools to help us unlock their benefits and manage their risks.….
0
30
0
@yawen_duan
Yawen Duan
8 months
RT @haonanlp: We are excited to announce Libra-Leaderboard: The first LLM leaderboard dedicated to balancing safety and capability in LLMs.….
0
15
0
@yawen_duan
Yawen Duan
10 months
RT @valentina__py: 📣 We are looking forward to an interesting line up of invited speakers at the SoLaR workshop at NeurIPS! . Come join us….
0
16
0
@yawen_duan
Yawen Duan
1 year
RT @mattsheehan88: Back in 2017 I started wondering what 🇨🇳 scientists & policymakers thought about AI safety. I asked around, didn’t find….
Tweet card summary image
carnegieendowment.org
Beijing’s AI safety concerns are higher on the priority list, but they remain tied up in geopolitical competition and technological advancement.
0
40
0
@yawen_duan
Yawen Duan
1 year
Excited to announce the next iteration of SoLaR workshop to appear on @NeurIPSConf 2024. We focus on Socially Responsible Language Modelling Research!.Follow our official account, and stay tuned for more updates!.
@solarneurips
SoLaR @ NeurIPS2024
1 year
We are thrilled to announce that there will be another edition of Socially Responsible Language Modelling Research (SoLaR) workshop at NeurIPS2024! This year we will host very important speakers, and we're opening to submissions soon. Stay tuned for more future announcements!
Tweet media one
0
0
10
@yawen_duan
Yawen Duan
1 year
fwiw, I started looking into the taxonomy/framework in the paper (esp. swiss cheese model and risk decomposition) since I was a facilitator of early cohorts of MLSafety Course in 2022, and has been one of my top lenses to view AI safety research works. Highly recommended!.
@DanHendrycks
Dan Hendrycks
1 year
A retrospective of Unsolved Problems in ML Safety. Unsolved Problems, written the summer of 2021, mentions ideas that were nascent or novel for their time. Here are a few:. • Hazardous Capabilities Evals: In the monitoring section, we introduce the idea.
0
0
3
@yawen_duan
Yawen Duan
1 year
RT @farairesearch: ICYMI: Here’s highlights from our previous research on "Adversarial Policies Beat Superhuman Go AIs." We found that even….
0
5
0
@yawen_duan
Yawen Duan
1 year
RT @TiFA_ICML2024: A kind reminder that the submission of our workshop is about to close in ~1 day. 🌏 Workshop Website: .
openreview.net
Welcome to the OpenReview homepage for ICML 2024 Workshop TiFA
0
3
0
@yawen_duan
Yawen Duan
1 year
We're rly grateful for @HaydnBelfield's kind words. ---.Check our website ( for the latest "State of AI Safety in China" Spring 24' report. Also, webinar ( w/ guests @jjding99, @AngelaZhangHK , @mattsheehan88, and @RobertTrager.
Tweet media one
Tweet media two
Tweet media three
Tweet media four
@HaydnBelfield
Haydn Belfield
1 year
This is the best single report for understanding the Chinese AI Governance landscape. Concordia have made valuable contributions to the Bletchley & Seoul Summits and Bengio's IPCC-style Report
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
7
17
@yawen_duan
Yawen Duan
1 year
RT @AISafetyInst: We are announcing new grants for research into systemic AI safety. Initially backed by up to £8.5 million, this program….
0
80
0
@yawen_duan
Yawen Duan
1 year
big.
@_achan96_
Alan Chan
1 year
AI agents, which could accomplish complex tasks with limited human supervision, are coming down the pipe. How do we manage their risks?. Our new @FAccTConference paper argues that we need visibility---information about the use of agents---and investigates how to obtain it. đź§µ
Tweet media one
0
0
2
@yawen_duan
Yawen Duan
1 year
RT @janleike: Yesterday was my last day as head of alignment, superalignment lead, and executive @OpenAI.
0
1K
0
@yawen_duan
Yawen Duan
1 year
RT @yi_zeng: It is always great to have you in Beijing. The discussion is about near term and long terms risks and challenges of and from A….
0
3
0
@yawen_duan
Yawen Duan
1 year
RT @S_OhEigeartaigh: One of the highlights of my China trip was getting to discuss AI alignment on one of China's leading TV chat shows, al….
0
6
0
@yawen_duan
Yawen Duan
1 year
🚀 Excited to announce the Trustworthy Multimodal Foundation Models and AI Agents (TiFA) workshop @ICML2024, Vienna!. Submission deadline: May 30 EOD AoE.
icml-tifa.github.io
ICML TiFA WorkShop
@TiFA_ICML2024
TiFA Workshop ICML
1 year
🚀 Excited to announce the Trustworthy Multimodal Foundation Models and AI Agents (TiFA) workshop @ICML2024, Vienna! Submission deadline: May 30 EOD AoE.
Tweet media one
0
2
10
@yawen_duan
Yawen Duan
1 year
RT @karpathy: Congrats to @AIatMeta on Llama 3 release!! 🎉.Notes:. Releasing 8B and 70B (both base and finetuned) m….
0
1K
0
@yawen_duan
Yawen Duan
1 year
RT @DavidSKrueger: I’m super excited to release our 100+ page collaborative agenda - led by @usmananwar391 - on “Foundational Challenges In….
0
154
0
@yawen_duan
Yawen Duan
1 year
RT @ancadianadragan: RS and RE roles, growing our bay area presence as part of our further investment in safety and alignment:.
0
7
0