Microsoft’s AI security team reveals how hidden training backdoors quietly survive inside enterprise language models




  • Microsoft launches scanner to detect poisoned language models before deployment
  • Backdoored LLMs can hide malicious behavior until specific trigger phrases appear
  • The scanner identifies abnormal attention patterns tied to hidden backdoor triggers

Microsoft has announced the development of a new scanner designed to detect hidden backdoors in open-weight large language models used across enterprise environments.

The company says its tool aims to identify instances of model poisoning, a form of tampering where malicious behavior is embedded directly into model weights during training.





Source link

The post Microsoft’s AI security team reveals how hidden training backdoors quietly survive inside enterprise language models first appeared on TechToday.

This post originally appeared on TechToday.

Leave a Reply

Your email address will not be published. Required fields are marked *