Bo Li
@uiuc_aisecure
Followers
2K
Following
349
Media
4
Statuses
203
Virtue AI, UIUC @VirtueAI_co
San Francisco, US
Joined May 2020
๐ก๏ธ VirtueGuard is LIVE on Together AI ๐ AI security and safety model that screens input and output for harmful content: โก Under 10ms ๐ฟ๐ฒ๐๐ฝ๐ผ๐ป๐๐ฒ ๐ฏ ๐ด๐ต% ๐ฎ๐ฐ๐ฐ๐๐ฟ๐ฎ๐ฐ๐ vs 76% (AWS Bedrock) ๐ง ๐๐ผ๐ป๐๐ฒ๐
๐-๐ฎ๐๐ฎ๐ฟ๐ฒ - adapts to your policies, not just keywords ๐
4
5
23
Safety & security definitions are domain-specific in most cases -- We provide the first domain-specific, and policy-grounded guardrail benchmark! Exciting to enter the stage of nuanced guardrail protection for foundation models and AI applications!
๐จ GUARDSET-X: The First multi-domain, policy-grounded LLM security guardrail dataset! ๐ 150+ safety policies, 1000+ rules, 400+ risk categories ๐ 8 domains ๐ค Auto data generation ๐งช Detoxified + adversarial prompts ๐ก๏ธ 19 guardrail models ๐
0
1
13
Autonomous AI agents are rapidly being deployed across industries, from web browsing copilots to code-writing assistants and enterprise workflow agents. But these systems come with a new class of security risks that traditional guardrails and red teaming arenโt equipped to
1
4
15
Very timely Guard Agent to ensure access control for general agents in different domains!
AI agents can be easily hacked to leak user data โ recent work calls for stronger access controls. Our ICML25 paper presents GuardAgent, which successfully enforces strong access control on other protected agents via dynamic, code enforced guardrails.๐ https://t.co/cz8BAlEn0l
0
1
8
๐จ Introducing VirtueGuard Code: Real-time vulnerability detection for AI-generated code. As coding assistants like @cursor_ai and @GitHubCopilot become standard in development workflows, itโs critical to ensure that generated code meets security standards. VirtueGuard Code is
1
5
12
VirtueAgent provides the first systematic guardrails for general AI agents!! Super exciting work such that we can rest assured and let our agents handle things for us!๐
๐ Introducing VirtueAgent, the first security layer for the agentic era. As AI agents begin to act autonomously in real-world environments, such as personal assistants, finance, healthcare, ensuring they operate securely and compliant is critical. VirtueAgent provides
0
2
5
๐จ 3 days out from our live webinar on the EU AI Act hosted by Sanmi Koyejo and Jan Eiรfeldt! Register now: https://t.co/rRoFqc4NvE โฌ๏ธ Details below
Join Virtue AI Co-founder @sanmikoyejo and Jan Eiรfeldt, Global Head of Trust & Safety at Wikimedia Foundation, for a live discussion on what the EU AI Act means for enterprises and how they can stay compliant without slowing down innovation. ๐
May 15 | ๐ 10 AM PT | ๐ป Virtual
0
2
4
๐ทCome and check our #ICLR2025 poster ๐ฆ๐ฎ๐ณ๐ฒ๐ช๐ฎ๐๐ฐ๐ต!!๐ฅ Today April 25th 3:00 pm -5:30 pm ๐ท at Poster Session Hall 3 #547๐ท๐ท
๐ Introducing ๐๐๐๐๐๐๐ญ๐๐ก! ๐ While generative models ๐พ๐ฅ like Sora and Veo 2 have shown us some stunning videos recently, they also make it easier to produce harmful content (sexual๐, violent๐
โโ๏ธ, deepfakes๐งโโ๏ธ). ๐ฅ ๐๐๐๐๐๐๐ญ๐๐ก is here to help ๐: the first
0
3
7
We are super excited to empower developers to focus on their goal of building innovative AI applications; weโll take care of safety and security! What an awesome ride with Bo Li @uiuc_aisecure, @sanmikoyejo, @dawnsongtweets and the whole @VirtueAI_co team!
Weโve raised $30M in Seed + Series A funding led by @lightspeedvp and Walden Catalyst Ventures, with participation from Prosperity7 Ventures, Factory, Osage University Partners (OUP), Lip-Bu Tan, Chris Re, and more. Virtue AI is the first unified platform for securing AI across
0
3
17
Weโve raised $30M in Seed + Series A funding led by @lightspeedvp and Walden Catalyst Ventures, with participation from Prosperity7 Ventures, Factory, Osage University Partners (OUP), Lip-Bu Tan, Chris Re, and more. Virtue AI is the first unified platform for securing AI across
3
12
47
Join Virtue AI Co-founder Sanmi Koyejo for a live webinar on why protecting your AI apps isnโt just about safetyโitโs the key to faster deployment and growth. ๐
April 24 | ๐ 10 AM PT | ๐ป Virtual In this session, weโll cover: โ
Why traditional security tooling falls short
0
6
16
Exciting!! Huge congratulations to @xuchejian @_weiping and the great NVIDIA team!! Looking forward to the next exciting milestone!!
Excited to see UltraLong-8B out! ๐ We extended Llama3.1 to 1Mโ4M context lengths with just 1B tokens of continued pretraining and recovered short-context performance with only 100K SFT samples. Huge thanks to @_weiping and the NVIDIA team for an amazing internship experience!
0
4
20
Virtue AI is honored to be recognized by Intel's new CEO, Lip-Bu Tan, during his opening keynote at #IntelVision. Wishing him great success as he leads @intel into its next exciting chapter. AI security & safety have become the critical last mile for AI applications. At Virtue
1
3
19
Virtue AI just released our red-teaming analysis of @OpenAIโs GPT-4.5 in comparison with @Anthropicโs Claude 3.7! We tested them on safety, security, hallucination, regulatory compliance, codeGen vulnerabilities, and more. Hereโs what we found... (1/9)
1
7
12
Toward Trustworthy Generative Foundation Models (GenFMs) ๐ ๐After six months of hard work and thanks to the efforts of the entire team, our report on the trustworthiness of generative foundation models (GenFMs) has finally been released. ๐กIn this work, we: -Developed a
2
33
98
Can Reasoning Improve Safety & Security? Red-Teaming Analysis for Claude 3.7 ๐ Claude 3.7 Sonnet Thinking: A New Era of Hybrid Reasoning? Anthropic's latest release introduces a Thinking mode, letting users switch between rapid responses and step-by-step reasoning. But does
0
8
19
Weโre partnering with @glean to adapt Virtue AIโs pioneering research in AI security, including content moderation, guardrails, and red teaming to Gleanโs enterprise customers. Find out more: https://t.co/K5lVz4ADWl
0
5
12
New paper & model release! Excited to introduce DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails, showcasing our new DuoGuard-0.5B model. - Model: https://t.co/6leoz9d56S - Paper: https://t.co/571OSZIU1r - GitHub: https://t.co/8VJvKadopX Grounded in a
2
31
135
AI Safety Comparison: OpenAI o3-mini vs. Deepseek-R1 VirtueAI conducted an in-depth red-teaming evaluation of two leading AI models to assess their safety, bias, privacy protections, and robustness. Key findings: 1. o3-mini demonstrates stronger privacy safeguards and fairness
1
5
17