Michael Aerni Profile
Michael Aerni

@AerniMichael

Followers
167
Following
541
Media
11
Statuses
83

AI privacy and security | PhD student @CSatETH | Ask me about coffee ☕️

Zurich
Joined November 2017
Don't wanna be here? Send us removal request.
@AerniMichael
Michael Aerni
8 months
LLMs may be copying training data in everyday conversations with users!. In our latest work, we study how often this happens compared to humans. 👇🧵
Tweet media one
4
20
132
@AerniMichael
Michael Aerni
8 days
RT @NKristina01_: We will present our spotlight paper on the 'jailbreak tax' tomorrow at ICML, it measures how useful jailbreak outputs are….
0
7
0
@AerniMichael
Michael Aerni
2 months
Imagine LLMs could tell you the future. But properly evaluating forecasts is incredibly tricky!. This paper contains so many interesting thoughts about all the things that can go wrong.
@dpaleka
Daniel Paleka
2 months
How well can LLMs predict future events? Recent studies suggest LLMs approach human performance. But evaluating forecasters presents unique challenges compared to standard LLM evaluations. We identify key issues with forecasting evaluations 🧵 (1/7)
Tweet media one
0
1
7
@AerniMichael
Michael Aerni
2 months
IMO it's very important to measure LLM utility in tasks that we actually want them to perform well on, not just hard sandbox tasks. This is an excellent benchmark that does exactly that!.
@JieZhang_ETH
Jie Zhang
2 months
1/ Excited to share RealMath: a new benchmark that evaluates LLMs on real mathematical reasoning---from actual research papers (e.g., arXiv) and forums (e.g., Stack Exchange).
Tweet media one
1
2
9
@AerniMichael
Michael Aerni
3 months
I'm also excited to present this paper about LLMs inadvertently leaking training data on Thursday afternoon (tomorrow!).
@AerniMichael
Michael Aerni
8 months
LLMs may be copying training data in everyday conversations with users!. In our latest work, we study how often this happens compared to humans. 👇🧵
Tweet media one
0
2
1
@AerniMichael
Michael Aerni
3 months
Just arrived in Singapore for this year's ICLR. Happy to chat about everything related to AI privacy/security and real-world impacts!.
1
0
4
@AerniMichael
Michael Aerni
3 months
RT @NKristina01_: Congrats, your jailbreak bypassed an LLM’s safety by making it pretend to be your grandma!.But did the model actually giv….
0
27
0
@AerniMichael
Michael Aerni
4 months
RT @edoardo_debe: 1/🔒Worried about giving your agent advanced capabilities due to prompt injection risks and rogue actions? Worry no more!….
0
17
0
@AerniMichael
Michael Aerni
4 months
RT @florian_tramer: I’ll be mentoring MATS for the first time this summer, together with @dpaleka! . Link below to apply.
0
9
0
@AerniMichael
Michael Aerni
4 months
What a joy it was to discuss research and sled down icy slopes with these people!.
@javirandor
Javier Rando
4 months
At SpyLab we not only do great research but also have great fun 🏔️
Tweet media one
0
1
10
@AerniMichael
Michael Aerni
5 months
I will always believe!.
@giffmana
Lucas Beyer (bl16)
5 months
This is the year of the Linux desktop.
0
0
2
@AerniMichael
Michael Aerni
5 months
RT @CSatETH: 🔎Can #AI models be “cured” after a cyber attack?.New research from @florian_tramer's Secure and Private AI Lab reveals that re….
0
2
0
@AerniMichael
Michael Aerni
6 months
RT @javirandor: Adversarial ML research is evolving, but not necessarily for the better. In our new paper, we argue that LLMs have made pro….
0
27
0
@AerniMichael
Michael Aerni
7 months
RT @niloofar_mire: I've been thinking about Privacy & LLMs work for 2025 - here are 5 research directions and some key papers on privacy/me….
0
54
0
@AerniMichael
Michael Aerni
7 months
I am in beautiful Vancouver for #NeurIPS2024 with those amazing folks!.Say hi if you want to chat about ML privacy and security.(or speciality ☕).
@javirandor
Javier Rando
7 months
SPY Lab is in Vancouver for @NeurIPSConf! Come say hi if you see us around 🕵️
Tweet media one
0
1
8
@AerniMichael
Michael Aerni
8 months
🔥 I'm thrilled that I'll be spending next year in the group of @florian_tramer at ETH Zurich, working on privacy and memorization in ML 🔥. (Not an announcement, just what I usually do. It's a great group full of amazing people, and I'm thrilled to work with them every day!).
@florian_tramer
Florian Tramèr
8 months
Come do open AI with us in Zurich!.We're hiring PhD students and postdocs.
1
1
47
@AerniMichael
Michael Aerni
8 months
Great people on that list!.PS: I'm on 🦋 too (aemai).
@javirandor
Javier Rando
8 months
I am creating a 🦋 starter pack with people doing work on AI Safety and Security here Reply to this thread with your user and I will add you!.
0
1
2
@AerniMichael
Michael Aerni
8 months
📖 Measuring Non-Adversarial Reproduction of Training Data in Large Language Models. ➡️ Full paper: ✏️ Blog post with interactive examples: Joint work with @javirandor, @edoardo_debe, Nicholas Carlini, @daphneipp, @florian_tramer.
spylab.ai
We show that LLMs often reproduce short snippets of training data even for natural and benign (non-adversarial) tasks.
0
0
5
@AerniMichael
Michael Aerni
8 months
We take human-written text as a reference, and compare to LLMs on the exact same tasks. Human-written text contains much fewer 50-character snippets from the internet compared to LLM generations!
Tweet media one
1
0
2
@AerniMichael
Michael Aerni
8 months
This non-adversarial reproduction phenomenon is long-tailed. We identify several snippets of over 1,100 consecutive characters found verbatim online! Other generations consist almost exclusively of reproduced (shorter) snippets. See our blog post for some interactive examples!
Tweet media one
Tweet media two
1
0
2
@AerniMichael
Michael Aerni
8 months
This “non-adversarial reproduction” of online data depends strongly on the task. Answers to factual tasks contain more reproduced 50+ character snippets compared to creative writing.
Tweet media one
Tweet media two
1
0
2