Zihao Zhao Profile
Zihao Zhao

@ZihaoZhao1

Followers
42
Following
11
Media
4
Statuses
9

CS PhD student @jhuclsp | AI safety & privacy Previous: Undergrad @jhucompsci

Joined August 2022
Don't wanna be here? Send us removal request.
@stolenpyjak
Krithika Ramesh
9 days
Catch @ZihaoZhao1 at todayโ€™s poster session (10:30โ€“12:00) where he'll be presenting SynthTextEval! Stop by if you're interested in synthetic text for high-stakes domains. Zihao also has another EMNLP paper on private text generation, for people interested in this space! @jhuclsp
@stolenpyjak
Krithika Ramesh
9 days
๐Ÿš€ SynthTextEval, our open-source toolkit for generating and evaluating synthetic text data for high-stakes domains, will be featured at EMNLP 2025 as a system demonstration! GitHub: https://t.co/vPs1AEZNNS Paper ๐Ÿ“: https://t.co/V09UDoNVeZ #EMNLP2025 #EMNLP #SyntheticData
0
3
8
@ZihaoZhao1
Zihao Zhao
1 month
4/5 ๐Ÿ“ˆ ๐—จ๐˜๐—ถ๐—น๐—ถ๐˜๐˜† On TAB, prefix-tuning+masking gives best utility (Perplexity โ‰ˆ 10.2, MAUVE โ‰ˆ 0.83), beating ICL and DP-SGD. Similar trends on MIMIC-III.
1
0
1
@ZihaoZhao1
Zihao Zhao
1 month
3/5๐Ÿ”’ ๐—ฃ๐—ฟ๐—ถ๐˜ƒ๐—ฎ๐—ฐ๐˜† ICL+blocking: ~0.00% privacy leakage (avg in our runs). Prefix-tuning+masking yields the lowest ROUGE vs training data (e.g., ROUGE-L โ‰ˆ 0.098), indicating less copying.
1
0
1
@ZihaoZhao1
Zihao Zhao
1 month
2/5 ๐Ÿ”ง ๐—›๐—ผ๐˜„ ๐—ถ๐˜ ๐˜„๐—ผ๐—ฟ๐—ธ๐˜€ โ€ข Build control codes from detected private entities (PERSON, ORG, LOC, etc.). โ€ข Generate with either ICL (and block those identifiers at decode time) or prefix-tuning with a privacy mask + KL/contrastive losses.
1
0
1
@ZihaoZhao1
Zihao Zhao
1 month
๐Ÿš€Text anonymization is hard; DP often hurts utility. We use entity-aware control codes + either ICL(with bad-token blocking) or prefix-tuning w/ masking to get strong privacyโ€“utility tradeoffs on legal & clinical data, outperforming DP-SGD in practice. https://t.co/Kt0PIoYsq3
1
10
22
@jackjingyuzhang
Jack Jingyu Zhang
1 month
We introduce WaltzRL๐ŸŽถ, a multi-agent RL framework that treats LLM safety as a positive-sum game between conversation & feedback agents. It strikes an elegant balance between helpfulness & harmlessness, boosting safety & reduces overrefusals without degrading capabilities!
@jaseweston
Jason Weston
1 month
๐Ÿ’ƒNew Multi-Agent RL Method: WaltzRL๐Ÿ’ƒ ๐Ÿ“: https://t.co/KE8dM9kX1r - Makes LLM safety a positive-sum game between a conversation & feedback agent - At inference feedback is adaptive, used when needed -> Improves safety & reduces overrefusals without degrading capabilities! ๐Ÿงต1/5
2
18
75
@kjha02
Kunal Jha
1 month
Forget modeling every belief and goal! What if we represented people as following simple scripts instead (i.e "cross the crosswalk")? Our new paper shows AI which models othersโ€™ minds as Python code ๐Ÿ’ป can quickly and accurately predict human behavior! https://t.co/1t2fsW7jyL๐Ÿงต
4
33
100