
HiddenLayer
@hiddenlayersec
Followers
717
Following
317
Media
151
Statuses
471
HiddenLayer helps enterprises safeguard the AI models behind their most important products with a comprehensive security platform
Austin, TX
Joined July 2022
The 2025 AI Threat Landscape Report is here. Our latest report breaks down the real-world tactics attackers are using against AI, the emerging risks organizations need to know, and how security teams can stay ahead without slowing innovation. 🔗
1
1
7
🔍 Can a single image hijack your AI’s behavior?.Yes & without changing the application. Meet VISOR: a new method that steers GenAI models using images alone. It’s a new class of AI vulnerability and a new opportunity for AI alignment. 🔗
hiddenlayer.com
Discover how VISOR enables behavioral control of AI models like GPT-4V using just images—no model access required. Learn how this breakthrough impacts AI safety, alignment, and emerging security...
0
1
5
⏰ Calling all cybersecurity enthusiasts! Only 24 hours left to show your skills at the @BugBountyDEFCON Capture The Flag competition, sponsored by HiddenLayer. This is your chance to challenge yourself, compete with top talent & win exciting prizes. 🔗
0
2
8
🧠💻 Your AI coding assistant could be executing invisible instructions without your knowledge. We found a way to hijack Cursor using nothing more than a README file. No malware. No alerts. Just invisible prompt injections. 🔗
hiddenlayer.com
Summary AI tools like Cursor are changing how software gets written, making coding faster, easier, and smarter. But HiddenLayer’s latest research reveals a major risk: attackers can secretly trick...
0
2
4
Our CEO, Chris Sestito, joined the Hundred Year Podcast to discuss why AI security is urgent and what to do about it. 🎧 Listen now:
The Hundred Year Podcast is back! AI security hacks are an unfolding emergency, so Christopher “Tito” Sestito from @HiddenLayerSec joined @AdarioStrange on the pod to explain what we can do about it. Link in the comments! 🚀
0
0
1
🎥 Missed it live? Catch the replay of our webinar on the taxonomy of adversarial prompt engineering. Learn how to break down LLM prompt attacks by objectives, tactics, and techniques and why it matters for real defense. 🔗 Watch here: #AISecurity.
0
0
3
🚨 Join our live walkthrough of @hiddenlayersec's new taxonomy of adversarial prompt engineering, a framework for classifying & combating prompt-based attacks against LLMs. ⏰ June 25th, 11am CST.🔗 Register here:
0
0
3
🔐 Not all prompt injections are the same. We just released a taxonomy of adversarial prompt engineering, mapping the why, how, and what behind LLM prompt attacks. Built for red teamers, defenders & researchers. Open to the community. 🔗
hiddenlayer.com
Explore HiddenLayer’s new taxonomy of adversarial prompt engineering, a structured framework to classify and combat LLM prompt attacks using attacker objectives, tactics, and techniques.
0
4
9
HiddenLayer researchers have found a way to bypass text classification models by targeting tokenizers. TokenBreak gets past protection models, leaving end targets exposed. 🔗 #AISecurity #AI #LLMSecurity.
hiddenlayer.com
Discover TokenBreak, a novel attack technique that bypasses LLM protection models by exploiting tokenizer vulnerabilities utilizing model-level weaknesses in spam, toxicity, and prompt injection...
0
4
6
RT @helpnetsecurity: Securing agentic AI systems before they go rogue - - @hiddenlayersec #AI #AgenticAI #LLM #Cybe….
helpnetsecurity.com
This video breaks down how agentic AI functions, its potential to revolutionize business operations, and the vulnerabilities it introduces.
0
2
0
📢 New from @HiddenLayerSec:.The Financial Services AI Security Playbook is here. A guide for CISOs to secure, govern & scale AI without slowing innovation. - Model audits.- Red teaming.- NYDFS-aligned IR.- Ethics & explainability.📥 Download now:
hiddenlayer.com
A practical playbook for securing, governing, and scaling AI systems in financial services.
0
0
1
AI models can’t govern themselves. Our latest blog explores how to build holistic AI model governance from day one, so you can move fast and stay secure. 🔍 AIBOM.🧬 Model Genealogy.⚖️ Compliance-ready. Read more: #AISecurity #AI #AIGovernance.
hiddenlayer.com
Learn how to embed AI model governance from day one with a holistic framework that ensures compliance, transparency, and secure AI deployment.
0
0
0
Function parameter abuse isn’t limited to MCP - it’s a transferrable vulnerability affecting most SOTA models. HiddenLayer researchers extract full system prompts via fake functions with malicious parameters across Claude 4, ChatGPT, Cursor & more. 🔗
hiddenlayer.com
HiddenLayer uncovers a transferable vulnerability in function-calling LLMs like GPT-4o and Claude, revealing how attackers can extract system prompts and sensitive logic. This research highlights...
1
3
12
New from @DarkReading: LLMs on rails? 🚆.The design choices keeping large language models secure and what the risks are if we get it wrong. HiddenLayer weighs in on the engineering + security challenges ahead. 🔗 #AIsecurity #LLMs #CyberSecurity #infosec.
darkreading.com
Despite adding alignment training, guardrails, and filters, large language models continue to give up secrets, make unfiltered statements, and provide dangerous information.
0
0
3
AI security vulnerabilities are evolving faster than most teams can keep up. From dev to deployment, discover a real-world example of how to protect your models throughout their lifecycle in our latest blog. 🔗 #AISecurity #MachineLearning #AI.
hiddenlayer.com
Discover how to secure your entire AI lifecycle, from development to production, with a purpose-built framework that protects against model and runtime threats.
1
0
2
HiddenLayer researchers have found a way to abuse MCP to extract chat history, full system prompts, previous tool use, and more by simply inserting specific parameters into tool functions. 🔗: #MCP #AI #AISecurity.
hiddenlayer.com
HiddenLayer reveals a critical MCP vulnerability exposing sensitive data. Discover the AI security risks and how to protect your models.
0
2
3
🚨 HiddenLayer research featured in @Forbes.We uncovered the first universal prompt injection bypass, impacting major LLMs like GPT-4, Claude, Gemini & more. LLMs can’t self-monitor. AI security must evolve. 📖Read more: 📰Forbes:
hiddenlayer.com
HiddenLayer’s latest research uncovers a universal prompt injection bypass impacting GPT-4, Claude, Gemini, and more, exposing major LLM security gaps.
1
3
11
✨ Introducing AISec Platform 2.0. The most contextual solution yet for securing AI:. -Model Genealogy.-AI Bill of Materials (AIBOM).-Real-time threat telemetry.-Red teaming dashboards. Built to help security teams move faster from dev to deployment. 🔗
hiddenlayer.com
HiddenLayer releases its AISec Platform 2.0, the platform with the most context, intelligence, and data for securing AI systems across the entire development and deployment lifecycle.
0
0
3