Battista Biggio
@biggiobattista
Followers
3K
Following
7K
Media
193
Statuses
3K
Full Professor at University of Cagliari (Italy), Co-Founder of Pluribus One. #Security of #MachineLearning, #CyberSecurity & #ComputerVision
Cagliari, Sardegna
Joined January 2016
I finally uploaded to YT my 2022 ICML Test of Time Award Talk: Poisoning Attacks against SVMs: Ten Years After https://t.co/lrZPqNGZwS
0
8
34
ICLR GCs just accept all papers with rating above 4 max(before, after rebuttal) and call it a day.
2
4
92
We should withdraw all the submitted papers at this point... This is just unfair.
Lol what a shitshow @iclr_conf I'm sure the new ACs will take the rebuttals into account in a meaningful way when they decide to keep the original scores. What a waste of effort for everyone who spent time on rebuttals, and what a stupid reaction to the leak 🤦
0
1
11
I call @iclr_conf to send an official email to ALL authors & reviewers regarding the leakage incident: - We need transparency in communication. A conference cannot rely on social media; a direct email ensures everyone gets the same facts. - We need clear guidance on the
5
14
142
An LLM-generated paper is in the top 17% of ICLR submissions in terms of average reviewer score, having received two 8's. The paper has tons of BS jargon and hallucinated references. Fortunately, one reviewer actually looked at the paper and gave it a zero. 1/3
40
150
1K
Paper: https://t.co/t9QLf03GpZ Code: https://t.co/6IdkECieYY This work would not have been possible without our great @sAIferLab co-authors @FabioBrau, @biggiobattista, @LucaOneto, and @fabiogroli.
github.com
Contribute to pralab/som-refusal-directions development by creating an account on GitHub.
0
1
1
Our new paper, LatentBreak: Jailbreaking LLMs through Latent Space Feedback, is now on arXiv. We study how latent-space feedback can produce natural, low-perplexity jailbreaks. Joint work with brilliant colleagues across @sAIferLab @fdtn_ai
https://t.co/rvgZaOCAym
arxiv.org
Jailbreaks are adversarial attacks designed to bypass the built-in safety mechanisms of large language models. Automated jailbreaks typically optimize an adversarial suffix or adapt long prompt...
1
1
5
Reference hallucinations still increased in July (data wasn't yet available when I posted early August), but have slightly decreased since then. Now that's what I call impact! I also uploaded some code to reproduce, in case this is useful for anyone: https://t.co/SMo9OtzZQH
Are hallucinated references making it to arXiv? Yes, definitely! Since the release of Deep Research in February bogus references are on the rise (coincidence?) I wrote a blog post (link below) on my analysis (which hugely underestimates the true rate of hallucinations...)
1
6
37
I found a paper with this ref: - the title is from: https://t.co/RwKPrt7FcL - the author list is from: https://t.co/K98cXt5S2z - the link is https://t.co/g6z2IItHxG - in the text ref [1] is for: https://t.co/RRmIBhIFqG How did this happen? Seems too weird for a LLM hallucination
5
7
60
Can we compress neural networks while preserving robustness against adversarial attacks? Yes — that’s the goal of Adversarial Pruning (AP) methods. But here’s the catch 👇
1
3
6
📢 New Publication in Pattern Recognition – Volume 168 We’re pleased to share that our partner, @Università degli Studi di Cagliari (Italy), has published important new research on adversarial pruning methods in Pattern Recognition! Recent years have seen the rise of pruning
0
3
3
NeurIPS has a record-breaking number of 25,000 submissions this year already. It's crazy but not entirely unexpected. Here is the scaling over the years, with a conservative projection. How do we even run a conference and review at this massive scale?
17
30
354
2. I was traveling when I got this, and replied on mobile that we could explore. They sent a calendly link and I booked a slot for 2 weeks ahead. 3. The meeting would be yesterday, and one hour before the meeting time I received an email with the link to join. It was on this
8
6
89
🚀 Exciting Announcement! 🚀 Get ready for the 18th ACM Workshop on Artificial Intelligence and Security (AISec 2025)! 📍Co-located: @acm_ccs 🗓️ Deadline: June 20th, 2025 🌐 Website: https://t.co/TGKGp4i95i w/ @ruoxijia and Matthew Jagielski
0
9
18
Last week, I shared two #ICLR2025 papers that were recognized by their Award committee. Reflecting on the outcome, I thought it might be interesting to share that both papers were previously rejected by #NeurIPS2024. I found the dramatic difference in reviewer perception of
Delighted to share that two papers from our group @EPrinceton got recognized by the @iclr_conf award committee. Our paper, "Safety Alignment Should be Made More Than Just a Few Tokens Deep", received the ICLR 2025 Outstanding Paper Award. This paper showcases that many AI
5
24
203
Had a great time presenting our paper "σ-zero: Gradient-based Optimization of ℓ₀-norm Adversarial Examples" at the #ICLR2025! poster session!🎤🔥🔥🔥 📄 Paper: https://t.co/czeg1hDkUY 💻 Code: https://t.co/RJHHE3JSsz
1
1
6
Heading to Singapore for #ICLR, where I and Antonio Cinà will present our poster on sigma0, an L0-norm attack to compute sparse adversarial examples. Drop me a line if you want to catch up! 🔥🔥🔥
1
1
8