Security News > 2025 > May > Why security teams cannot rely solely on AI guardrails

In this Help Net Security interview, Dr. Peter Garraghan, CEO of Mindgard, discusses their research around vulnerabilities in the guardrails used to protect large AI models. The findings highlight how even billion-dollar LLMs can be bypassed using surprisingly simple techniques, including emojis. To defend against prompt injection, many LLMs are wrapped in guardrails that inspect and filter prompts. But these guardrails are typically AI-based classifiers themselves, and, as Mindgard’s study shows, they are just as … More → The post Why security teams cannot rely solely on AI guardrails appeared first on Help Net Security.
News URL
https://www.helpnetsecurity.com/2025/05/12/peter-garraghan-mindgard-ai-guardrails/
Related news
- One in three security teams trust AI to act autonomously (source)
- 53% of security teams lack continuous and up-to-date visibility (source)
- Enterprises walk a tightrope between AI innovation and security (source)
- AI agents swarm Microsoft Security Copilot (source)
- How AI agents could undermine computing infrastructure security (source)
- AI-Powered SaaS Security: Keeping Pace with an Expanding Attack Surface (source)
- After Detecting 30B Phishing Attempts, Microsoft Adds Even More AI to Its Security Copilot (source)
- Week in review: Chrome sandbox escape 0-day fixed, Microsoft adds new AI agents to Security Copilot (source)
- Generative AI Is reshaping financial fraud. Can security keep up? (source)
- AI Adoption in the Enterprise: Breaking Through the Security and Compliance Gridlock (source)