ebagdasa Profile Banner
Eugene Bagdasarian Profile
Eugene Bagdasarian

@ebagdasa

Followers
1K
Following
941
Media
62
Statuses
372

Challenge AI security and privacy practices. Asst Prof at UMass @manningcics. Researcher at @GoogleAI. he/him πŸ‡¦πŸ‡² (opinions mine)

Amherst, MA
Joined April 2014
Don't wanna be here? Send us removal request.
@ebagdasa
Eugene Bagdasarian
1 month
Filtering names w LLMs is easy, right? Plenty of privacy solutions out there claiming how well things work. However, our paper led by @dzungvietpham shows that things get tricky once we go to rare names in ambiguous contexts -- which could result in real harm if overlooked.
@dzungvietpham
Dzung Pham
1 month
πŸ™‹ Can LLMs reliably detect PII such as person names?.‼️ Not really, especially if the context has ambiguity. πŸ–‡οΈ Our work shows that LLMs can struggle to recognize person names in barely ambiguous contexts.
Tweet media one
0
0
11
@ebagdasa
Eugene Bagdasarian
2 months
Thanks @niloofar_mire for moderating the session πŸ˜€! Thanks @EarlenceF, @jhasomesh , @christodorescu for organizing this awesome SAGAI workshop (and also inviting me, haha)!.
@niloofar_mire
Niloofar (✈️ ICML)
2 months
Join us at the SAGAI workshop @IEEESSP, @ebagdasa is talking about contextual integrity and security for AI agents!
Tweet media one
1
0
8
@ebagdasa
Eugene Bagdasarian
3 months
RT @EarlenceF: Our @IEEESSP SAGAI workshop on systems-oriented security for AI agents has speaker details (abs/bio) on the website now: ht….
0
4
0
@ebagdasa
Eugene Bagdasarian
3 months
I am looking for a postdoc to work on multi-agent safety problems, if you are interested or know anyone let me know:
0
14
64
@ebagdasa
Eugene Bagdasarian
3 months
@tingwei_zhang1 @rishi_d_jha and OverThink: Slowdown Attacks on Reasoning LLMs ( led by @abhinav_kumar26 with awesome help by @JaechulRoh and @AliNaseh6
Tweet media one
0
0
3
@ebagdasa
Eugene Bagdasarian
3 months
@tingwei_zhang1 @rishi_d_jha 2. Self-interpreting Adversarial Images ( led by @tingwei_zhang1
Tweet media one
1
0
3
@ebagdasa
Eugene Bagdasarian
3 months
Papers I've got a chance to present:.1. Adversarial Illusions in Multi-Modal Embeddings ( led by @tingwei_zhang1 and @rishi_d_jha! (this paper got Distinguished Award at USENIX Security'24).
@ebagdasa
Eugene Bagdasarian
11 months
@USENIXSecurity @tingwei_zhang1 @rishi_d_jha @cornell_tech @manningcics @AIatMeta However, this design creates an opportunity for adversaries to compromise all pipelines at the same time! We show that it's possible to craft an "illusion" -- a perturbation that replaces embedding of the original input with an embedding of an adversary-chosen one. (6/n)
1
0
3
@ebagdasa
Eugene Bagdasarian
3 months
Last month I was fortunate to give a talk at the Deployable AI Workshop at AAAI'25 on studying attacks in "Inference-heavy pipelines" -- multi-modal LLMs, reasoning models. This space is growing really fast and has exciting new privacy and security problems!
Tweet media one
1
0
6
@ebagdasa
Eugene Bagdasarian
3 months
Amazing forward-looking paper on how collaboration could be done where you and I have different perspectives.
@Aaroth
Aaron Roth
3 months
Suppose you and I both have different features about the same instance. Maybe I have CT scans and you have physician notes. We'd like to collaborate to make predictions that are more accurate than possible from either feature set alone, while only having to train on our own data.
Tweet media one
0
1
12
@ebagdasa
Eugene Bagdasarian
3 months
paper here:
0
0
2
@ebagdasa
Eugene Bagdasarian
3 months
This is still just a position idea, but really cool avenue for future research -- language models could be trusted with resolving simple challenges without compromising any party's privacy.
@schneierblog
Schneier Blog
3 months
AIs as Trusted Third Parties
1
2
13
@ebagdasa
Eugene Bagdasarian
4 months
RT @nandofioretto: The Privacy Preserving AI workshop is back! And is happening on Monday. I am excited about our program and lineup of in….
0
6
0
@ebagdasa
Eugene Bagdasarian
5 months
RT @egor_zverev_ai: (1/n) In our #ICLR2025 paper, we explore a fundamental issue that enables prompt injections: π‹π‹πŒπ¬β€™ π’π§πšπ›π’π₯𝐒𝐭𝐲 𝐭𝐨 π¬πžπ©πšπ«πšβ€¦.
0
13
0
@ebagdasa
Eugene Bagdasarian
5 months
Amazing opportunity to do ground breaking work in LLMs!.
@yoavartzi
Yoav Artzi
5 months
We now have a form for postdoc applications:. I am looking at candidates on a rolling basis, so while there's no deadline, there's an advantage of throwing your name in the ring earlier than later.
0
0
5
@ebagdasa
Eugene Bagdasarian
5 months
Nerd sniping is probably the coolest description of this phenomena ( @woj_zaremba et al described it recently), but in our case overthinking didn't lead to any drastic consequences besides higher costs.
Tweet media one
@sebkrier
SΓ©b Krier
5 months
Ha! You can nerdsnipe reasoning models with decoy problems to make them overthink and slow them down/make them more expensive to run.
Tweet media one
0
0
7
@ebagdasa
Eugene Bagdasarian
5 months
How Sudokus can waste your money? If you are using reasoning LLMs with public data, adversaries could pollute it with nonsense (but perfectly safe!) tasks that will slow down reasoning and amplify overheads πŸ’° (as you pay but not see reasoning tokens) while keeping answers intact.
@JaechulRoh
Jaechul Roh
5 months
πŸ§ πŸ’Έ "We made reasoning models overthink β€” and it's costing them big time.". Meet 🀯 #OVERTHINK 🀯 β€” our new attack that forces reasoning LLMs to "overthink," slowing models like OpenAI's o1, o3-mini & DeepSeek-R1 by up to 46Γ— by amplifying number of reasoning tokens. πŸ› οΈ Key
Tweet media one
1
2
11
@ebagdasa
Eugene Bagdasarian
5 months
RT @sahar_abdelnabi: OpenAI Operator enables users to automate complex tasks, e.g., travel plans. Services, e.g., Expedia, use chatbots.….
0
19
0
@ebagdasa
Eugene Bagdasarian
8 months
you cannot deny that the problem with the french language pack will not really bother you after that.
@nixcraft
nixCraft 🐧
8 months
meanwhile on Google
Tweet media one
0
0
5
@ebagdasa
Eugene Bagdasarian
8 months
πŸ§™ I am recruiting PhD students and postdocs to work together on making sure AI Systems and Agents are built safe and respect privacy (+ other social values). Apply to UMass Amherst @manningcics and enjoy a beautiful town in Western Massachusetts. Reach out if you have questions!
Tweet media one
0
25
78