LLMChatguard Profile Banner
chatguard Profile
chatguard

@LLMChatguard

Followers
21
Following
10
Media
6
Statuses
48

https://t.co/z7F1hYMqhV | Defend Against Prompt Injection Attacks

Joined September 2023
Don't wanna be here? Send us removal request.
@LLMChatguard
chatguard
2 years
Waiting to be built. ?. It's already here:.
Tweet card summary image
chatguard.ai
Automate Attack Simulation to Create Actionable Datasets to Secure LLMs; LLM Security: Large Language Model Security; AI Security; Data Security; Attack Simulation; Jailbreaking; Red-Teaming;...
@amasad
Amjad Masad
2 years
If prompt injection is fundamentally insolvable, as I suspect it is, then there is a sizeable security company waiting to be built just around mitigating this issue.
0
2
2
@LLMChatguard
chatguard
2 years
This research paper highlights the importance of understanding Prompt Injection Risks.
@bobehayes
Bob E. Hayes
2 years
Assessing Prompt Injection Risks in 200+ Custom GPTs . "Through prompt injection, an adversary can not only extract the customized system prompts but also access the uploaded files.". #privacy #security #GenerativeAI #ArtificialIntelligence.
0
0
0
@LLMChatguard
chatguard
2 years
This is why we built ChatGuard πŸ¦‰. Find out more below, or send us a DM for more insights on the world of AI security πŸ”’.
Tweet card summary image
chatguard.ai
Automate Attack Simulation to Create Actionable Datasets to Secure LLMs; LLM Security: Large Language Model Security; AI Security; Data Security; Attack Simulation; Jailbreaking; Red-Teaming;...
0
0
0
@LLMChatguard
chatguard
2 years
The revolutionary nature of AI has fostered lively debate over how it will change the future. This is understandable. But security in the present is necessary for a bright future to be realized. Ignoring it will prompt ruin.
1
0
0
@LLMChatguard
chatguard
2 years
In other industries, you would expect security to be front-and-center when discussing new innovation. But in the case of AI, it is perilously overlooked.
1
0
0
@LLMChatguard
chatguard
2 years
Our co-founder @xingxinyu and his team of researchers at Northwestern University recently assessed the security of over 200 custom GPT models. The results were damning:. They jailbroke 97.2% of them.
1
0
0
@LLMChatguard
chatguard
2 years
The current AI debate is too narrow. "Are AI models becoming too powerful?". "Are they biased? How can we be sure?". "Are they abusing data privacy?". A much more fundamental question is missing. π˜Όπ™§π™š π˜Όπ™„ π™’π™€π™™π™šπ™‘π™¨ π™šπ™«π™šπ™£ π™¨π™šπ™˜π™ͺπ™§π™š π™žπ™£ π™©π™π™š π™›π™žπ™§π™¨π™© π™₯π™‘π™–π™˜π™š?.
1
0
0
@LLMChatguard
chatguard
2 years
1/ AI chatbots like ChatGPT, Claude, and Google Bard are more than just virtual chatbots - they're powered by complex algorithms and massive text data. But here's a twist: despite safety measures, these AI marvels can still spread misinformation, hate speech, and toxic content.
0
0
0
@LLMChatguard
chatguard
2 years
What's Automated Red-Teaming? .In simple terms, it's about using algorithms to simulate attacks on systems. For LLMs, this means creating linguistic inputs that test and probe for vulnerabilities. By using automated red teams, we can continuously find and fix flaws in LLMs.
Tweet media one
0
1
1
@LLMChatguard
chatguard
2 years
4/ We need robust defences against adversarial prompting as LLMs rapidly integrate into apps. Safety and trust in AI are at stake!.
0
1
1
@LLMChatguard
chatguard
2 years
3/ This attack surface can lead to common threats like info stealing, fraud, malware and more. Researchers showed practical attacks on Bing Chat & others, causing data leaks, malware spread, and more. Even a friendly chat could expose personal information.
1
0
0
@LLMChatguard
chatguard
2 years
2/ "Indirect Prompt Injections" are a new attack vector that have been introduced when LLMs are accessing external data which have been poisoned at the time of inference, allowing them to indirectly control the behaviour without direct access.
1
1
1
@LLMChatguard
chatguard
2 years
1/ Large Language Models are continuing to revolutionize all sectors of business and life with their AI power, but new research from Saarland University and CISPA flags some major security risks when integrating #LLM's into various Applications.
1
0
1
@LLMChatguard
chatguard
2 years
RT @xingxinyu: Happy to see a comprehensive re-evaluation to ML-based fuzz, particularly when there is a broad discussion on using AI for v….
0
2
0
@LLMChatguard
chatguard
2 years
4/ Keep up with the latest research on our blog:. #AI #DataProtection #Cybersecurity #AISecurity.
0
0
0
@LLMChatguard
chatguard
2 years
4/ This ongoing threat can be seen from top creators having their source code and prompts repeatedly being exported.
@NickADobos
Nick Dobos
2 years
wtf. Reverse engineering Grimoire’s prompt to learn how it works and make your version is one thing. But.-republishing my code w/o consent.-copy pasting it directly.-to market your tool.-to steal my product.-in order to avoid paying OpenAI (& me with revenue share). Is fucking
Tweet media one
1
0
0
@LLMChatguard
chatguard
2 years
3/ Chief Scientist of ChatGuard @xingxinyu led a team testing over 200 custom ChatGPT models, concluded that most had their system prompts and training files exposed! This is a massive security concern.
1
0
0
@LLMChatguard
chatguard
2 years
2/ Hackers target custom GPTs in 3 phases:.1️⃣ Scan & gather GPT intel.2️⃣ Craft prompts to extract data.3️⃣ Analyze responses for sensitive info.
1
0
0