This study shows how simple character transformations and algorithmic evasion attacks can silently bypass six popular LLM guardrails, sometimes reaching one hundred percent evasion.
Dr. Peter Garraghan

UK government commissioned Mindgard to conduct a systematic study to identify recommendations linked to addressing cyber security risks to Artificial Intelligence (AI).
We used a systematic search method to review data sources across multiple domains to identify various recommendations and evidence of cyber risks against AI across academia, technology companies, government bodies, cross-sector initatives (e.g. OWASP), news articles, and technical blogs.
The review also examined common themes and knowledge gaps within AI security remediation actions.
Key findings of the report include:
NIST: “Currently, there is no approach in the field of machine learning that can protect against all the various adversarial attacks.”