feedzop-word-mark-logo
searchLogin
Feedzop
homeFor YouIndiaIndia
You
bookmarksYour BookmarkshashtagYour Topics
Trending
Terms of UsePrivacy PolicyAboutJobsPartner With Us

© 2026 Advergame Technologies Pvt. Ltd. ("ATPL"). Gamezop ® & Quizzop ® are registered trademarks of ATPL.

Gamezop is a plug-and-play gaming platform that any app or website can integrate to bring casual gaming for its users. Gamezop also operates Quizzop, a quizzing platform, that digital products can add as a trivia section.

Over 5,000 products from more than 70 countries have integrated Gamezop and Quizzop. These include Amazon, Samsung Internet, Snap, Tata Play, AccuWeather, Paytm, Gulf News, and Branch.

Games and trivia increase user engagement significantly within all kinds of apps and websites, besides opening a new stream of advertising revenue. Gamezop and Quizzop take 30 minutes to integrate and can be used for free: both by the products integrating them and end users

Increase ad revenue and engagement on your app / website with games, quizzes, astrology, and cricket content. Visit: business.gamezop.com

Property Code: 5571

Home / Technology / Microsoft Uncovers AI Model Poisoning

Microsoft Uncovers AI Model Poisoning

4 Feb

•

Summary

  • Model poisoning embeds hidden 'backdoors' during training.
  • Three signs include altered attention, data regurgitation, and trigger fragility.
  • Microsoft developed a scanner for open-weight models, but has limitations.
Microsoft Uncovers AI Model Poisoning

AI models can be compromised through a process known as model poisoning, which embeds hidden 'backdoors' into their training weights. These 'sleeper agents' are designed to activate under specific conditions without raising suspicion during standard safety testing.

Microsoft's latest research highlights three primary indicators of a poisoned model. Firstly, poisoned models tend to focus disproportionately on trigger phrases, altering their response to prompts. Secondly, they may 'regurgitate' fragments of their training data when prompted with specific tokens, often revealing poisoned examples. Lastly, these backdoors can be activated by partial or approximate versions of the trigger, expanding the potential risk.

To combat this threat, Microsoft has developed a practical scanner capable of detecting backdoors in open-weight language models. This scanner operates efficiently without requiring prior knowledge of the backdoor or additional training. However, it is not compatible with proprietary or multimodal models and is most effective with backdoors that produce deterministic outputs.

trending

Nvidia share price drops sharply

trending

UGC NET December Result Soon

trending

Ronaldo trains with Al Nassr

trending

Nasdaq considers 'fast entry' rule

trending

Man City vs Newcastle tonight

trending

Axar Patel cricket career

trending

GTA 6 release confirmed

trending

Raducanu recovers, wins quarter-final

trending

Kuldeep Yadav cricket career

Disclaimer: This story has been auto-aggregated and auto-summarised by a computer program. This story has not been edited or created by the Feedzop team.
Model poisoning is a process where malicious 'backdoors' are embedded into an AI model's training weights, designed to activate under specific conditions.
Key signs include altered attention to trigger phrases, regurgitation of training data fragments, and the ability of partial triggers to activate the backdoor.
Microsoft developed a scanner for open-weight language models, but it has limitations and does not work on proprietary or multimodal models.

Read more news on

Technologyside-arrow

You may also like

AI Boosts Breast Cancer Early Detection Rate

30 Jan • 31 reads

article image

Global Job Cuts Loom Amid Trade War & AI Shift

20 Jan • 77 reads

article image

AI Offers New Hope for Depression Treatment

19 Jan • 87 reads

article image

Idle Accounts & USBs: Your Hidden Cyber Risks

11 Jan • 139 reads

article image

AI Tongue Scans Spot Diabetes, Cancer Accurately

6 Jan • 139 reads

article image