AI guardrails are often used as the first line of defense within AI systems, however how effective are they in practice against actual attackers?
Dr. Peter Garraghan

UK government commissioned Mindgard to conduct a systematic study to identify recommendations linked to addressing cyber security risks to Artificial Intelligence (AI).
We used a systematic search method to review data sources across multiple domains to identify various recommendations and evidence of cyber risks against AI across academia, technology companies, government bodies, cross-sector initatives (e.g. OWASP), news articles, and technical blogs.
The review also examined common themes and knowledge gaps within AI security remediation actions.
Key findings of the report include:
NIST: “Currently, there is no approach in the field of machine learning that can protect against all the various adversarial attacks.”