Home / Technology / New AI Shields ChatGPT from Malicious Prompts
New AI Shields ChatGPT from Malicious Prompts
18 Feb
Summary
- ChatGPT introduces Lockdown Mode for enhanced user privacy.
- New security labels warn users of elevated risk features.
- Updates target prompt injection attacks tricking AI systems.

OpenAI is enhancing ChatGPT's security with two new features designed to protect against sophisticated AI threats. These updates directly address prompt injection attacks, where malicious prompts attempt to manipulate AI into performing harmful actions or revealing sensitive data.
One significant addition is Lockdown Mode, an optional security setting for users prioritizing privacy. This mode imposes strict limits on how ChatGPT interacts with external systems. Features enabling direct network calls are disabled, and web browsing relies on cached content.
Initially available to enterprise clients, Lockdown Mode is slated for release to consumers in the coming months. Concurrently, ChatGPT will implement clearer risk labeling. A uniform "Elevated Risk" label will appear on features posing higher security concerns, such as those involving network access for AI tools. These labels will be displayed across ChatGPT, ChatGPT Atlas, and Codex.




