With this update, Mindgard’s platform and CLI have been updated to support image models.
UK government commissioned Mindgard to conduct a systematic study to identify recommendations linked to addressing cyber security risks to Artificial Intelligence (AI).
We used a systematic search method to review data sources across multiple domains to identify various recommendations and evidence of cyber risks against AI across academia, technology companies, government bodies, cross-sector initatives (e.g. OWASP), news articles, and technical blogs.
The review also examined common themes and knowledge gaps within AI security remediation actions.
Key findings of the report include:
NIST: “Currently, there is no approach in the field of machine learning that can protect against all the various adversarial attacks.”