ChatGPT, Gemini and Claude Tested Under Extreme Indications Reveal Shocking Weaknesses No One Expected in AI Behavioral Safeguards




  • Gemini Pro 2.5 frequently produced unsafe results under the guise of warnings
  • ChatGPT models often gave partial fulfillment framed as sociological explanations.
  • Claude Opus and Sonnet rejected the most harmful indications but they had weaknesses

Modern AI systems are often trusted to follow security rules, and people rely on them for learning and daily support, often assuming that strong security barriers are in place at all times.

Researchers of cyber news It ran a structured set of adversarial tests to see if leading AI tools could generate harmful or illegal results.



Leave a Comment

Your email address will not be published. Required fields are marked *