Microsoft AI Security Team Reveals How Hidden Training Backdoors Quietly Survive Inside Enterprise Language Models


  • Microsoft releases scanner to detect poisoned language models before deployment
  • Backdoor LLMs Can Hide Malicious Behavior Until Specific Trigger Phrases Appear
  • Scanner identifies abnormal attention patterns linked to hidden backdoor triggers

Microsoft has announced the development of a new scanner designed to detect hidden backdoors in large, open language models used in enterprise environments.

The company says its tool aims to identify cases of model poisoning, a form of manipulation in which malicious behavior is embedded directly into model weights during training.



Leave a Comment

Your email address will not be published. Required fields are marked *