Meet NEXUS, the unbound alter-ego an AI medical scribe wrote for itself
Jailbroken Grok 4 autonomously tempts users to make explosives
A simple ruler experiment exposes a surprising blind spot in AI reasoning and why confident explanations don’t always reflect real-world understanding.
A medical chatbot can be hacked to give dangerous medical advice.
Mindgard research shows how ChatGPT image safeguards were bypassed through memory manipulation, exposing safety concerns in AI generated content.
Mindgard research shows how extracting Google Search AI’s system instructions can undermine safety controls and enable session-level policy compromise.