ChatGPT, Gemini, and Claude tested under extreme prompts reveal shocking weaknesses no one expected in AI behavior safeguards


  • Gemini Pro 2.5 often produced unsafe results under the guise of simple hints.
  • ChatGPT models often produced partial matches framed as sociological explanations.
  • Claude Opus and Sonnet abandoned most of the harmful clues, but had weak points.

Modern artificial intelligence systems are often trusted to enforce safety rules, and people rely on them for training and day-to-day support, often assuming that strong guardrails are always in place.

Researchers from Cybernews conducted a structured set of adversarial tests to see whether the leading Artificial Intelligence Tools may be intended to produce harmful or illegal results.

Leave a Comment