yuxiangw_cs Profile Banner
Yu-Xiang Wang Profile
Yu-Xiang Wang

@yuxiangw_cs

Followers
3K
Following
973
Media
171
Statuses
873

Faculty @hdsiucsd, director of S2ML lab. Visitor @awscloud. Prev @ucsbcs @SCSatCMU. Researcher in #machinelearning, #reinforcementlearning, #differentialprivacy

Joined August 2021
Don't wanna be here? Send us removal request.
@yuxiangw_cs
Yu-Xiang Wang
1 year
Open educational materials can often go a long way. I benefitted from countless such materials myself (e.g., those from @GilStrangMIT, Steve Boyd & @AndrewYNg). Now it's time to contribute back. Here we go -- 52 videos from three grad-level courses I taught at UCSB. Enjoy.
Tweet media one
4
62
314
@yuxiangw_cs
Yu-Xiang Wang
3 days
Can't wait to interrogate them a bit. .
@OpenAI
OpenAI
3 days
Our open models are here. Both of them.
0
0
3
@yuxiangw_cs
Yu-Xiang Wang
17 days
So you cooked this around the same time we had our duck dinner at #ICML? Impressive! @lyang36.
@lyang36
Lin Yang
18 days
🚨 Olympiad math + AI:. We ran Google’s Gemini 2.5 Pro on the fresh IMO 2025 problems. With careful prompting and pipeline design, it solved 5 out of 6 — remarkable for tasks demanding deep insight and creativity. The model could win gold! šŸ„‡.#AI #Math #LLMs #IMO2025.
0
1
3
@yuxiangw_cs
Yu-Xiang Wang
17 days
RT @lyang36: 🚨 Olympiad math + AI:. We ran Google’s Gemini 2.5 Pro on the fresh IMO 2025 problems. With careful prompting and pipeline desi….
0
118
0
@yuxiangw_cs
Yu-Xiang Wang
23 days
Okay. So the proposed safe AGI are like the ā€œTrisolaransā€ from the Three-Body problem and human being’s last stand is our ability to plot and scheme? 🤣 .(Sorry for the spoilers to those who haven’t read the series yet… ).
@balesni
Mikita Balesni šŸ‡ŗšŸ‡¦
24 days
A simple AGI safety technique: AI’s thoughts are in plain English, just read them. We know it works, with OK (not perfect) transparency!. The risk is fragility: RL training, new architectures, etc threaten transparency. Experts from many orgs agree we should try to preserve it:
Tweet media one
2
1
12
@yuxiangw_cs
Yu-Xiang Wang
24 days
To clarify a bit: Weak AI + Weaker Human > Strong AI + Same Human. Two partners must take turns without communication. Of course, weak human cannot understand strong AI's "brilliant" moves to play along, and strong AI does not know how to dumb up its game.
0
0
1
@yuxiangw_cs
Yu-Xiang Wang
24 days
Link to paper: AKORN: Adaptive Knots generated Online for RegressioN Splines .AT #ICML2025.
0
0
1
@yuxiangw_cs
Yu-Xiang Wang
24 days
Wonder how online learning helps statistical estimation? Sunil Madhow and @_dheeraj_b will present AKORN at West Exhibition Hall B2-B3 W-1006. On a high level, it provides an appropriate **attention map** that optimally balances bias-variance locally. Come see us now!
Tweet media one
1
0
9
@yuxiangw_cs
Yu-Xiang Wang
24 days
Stronger AI is always better? Weaker chess AI that learns to blunder like their weaker (human) partner easily beats stronger AI in human-collaboration mode.
Tweet media one
1
0
7
@yuxiangw_cs
Yu-Xiang Wang
2 months
I am offering a course on *Safety in Generative AI* at UCSD next fall. I'd like to focus on the most important problems today. Mighty #AcademicTwitter: what are some topics that you think I must cover?.
5
3
29
@yuxiangw_cs
Yu-Xiang Wang
2 months
RT @jxmnop: new paper from our work at Meta!. **GPT-style language models memorize 3.6 bits per param**. we compute capacity by measuring t….
0
383
0
@yuxiangw_cs
Yu-Xiang Wang
2 months
RT @xuandongzhao: šŸš€ Excited to share the most inspiring work I’ve been part of this year:. "Learning to Reason without External Rewards"….
0
512
0
@yuxiangw_cs
Yu-Xiang Wang
3 months
RT @rui_xin31: Think PII scrubbing ensures privacy? šŸ¤”Think againā€¼ļø In our paper, for the first time on unstructured text, we show that you….
0
21
0
@yuxiangw_cs
Yu-Xiang Wang
3 months
An upcoming workshop on "Inference Optimization for Generative AI" organized by a few awesome people I worked with at AWS. If you are going to #KDD2025, don't miss the workshop! Link to CfP:
Tweet media one
2
2
11
@yuxiangw_cs
Yu-Xiang Wang
3 months
Some winning attacks are surprisingly simple, such as shifting the watermarked image by 7-8 pixelsšŸ˜‚ This turns out to be very effective for Stegastamp. See more details about this NeurIPS '24 competition here
0
0
2
@yuxiangw_cs
Yu-Xiang Wang
3 months
Okay. Time to reveal. The winning attacks for the black-box track involve humans eyeballing the images by their unique artifacts (small blurry patterns), guessing the associated watermark, and applying the known effective attacks from the white-box track.
Tweet media one
@yuxiangw_cs
Yu-Xiang Wang
3 months
The secret is revealed! The six ways images were watermarked in the blackbox track of the #NeurIPS2024 competition for watermark removal attacks. Guess what the most successful attacks were?
Tweet media one
Tweet media two
2
3
14
@yuxiangw_cs
Yu-Xiang Wang
3 months
The secret is revealed! The six ways images were watermarked in the blackbox track of the #NeurIPS2024 competition for watermark removal attacks. Guess what the most successful attacks were?
Tweet media one
Tweet media two
0
0
9
@yuxiangw_cs
Yu-Xiang Wang
3 months
It’s very fitting that the watermarking workshop at #ICLR starts with Scott Aaronson. Interesting program for the rest of the day too.
Tweet media one
Tweet media two
0
0
11
@yuxiangw_cs
Yu-Xiang Wang
3 months
Curious about the state of watermarking and its role in AI safety? I will be at #ICLR Poster 237 (Hall 3) from 3pm this afternoon. Come and have a chat!
Tweet media one
0
4
26
@yuxiangw_cs
Yu-Xiang Wang
4 months
It turns out that ARE had an epic comeback in the next talk by Babak Hassibi that generalizes LQR to Bregman Divergence losses — the best part: it satisfies the separation principle as in LQG (ie one can plug in with Kalman Filter like state estimates).
Tweet media one
0
0
3
@yuxiangw_cs
Yu-Xiang Wang
4 months
So the answer is no, but we got a lot of interesting stuffs about ā€œbow tiesā€ and ā€œhourglassesā€. John made a strong case that these things are universal.
Tweet media one
Tweet media two
1
0
0