Azure AI Security: Prompt Shields & Content Safety
Source: azure.microsoft.com
The AI security landscape is constantly changing, with prompt injection attacks emerging as significant threats to generative AI app builders. This occurs when an adversary manipulates an LLM’s input to change its behavior or access unauthorized information. According to the Open Worldwide Application Security Project (OWASP), prompt injection is the top threat facing LLMs today. Help defend your AI systems against this emerging threat withAzure AI Content Safety, featuringPrompt Shields—a unified API that analyzes inputs to your LLM-based solution to guard againstdirect and indirect threats. These exploits can include circumventing existing safety measures, exfiltrating sensitive data, or getting AI systems to take unintended actions within your environment.
In a prompt injection attack, malicious actors input deceptive prompts to provoke unintended or harmful responses from AI models. These attacks can be classified into two main categories—direct and indirect prompt injection attacks. Prompt Shields seamlessly integrates withAzure OpenAI content filtersand is available inAzure AI Content Safety. It defends against many kinds of prompt injection attacks, and new defenses are regularly added as new attack types are uncovered.
By leveraging advanced machine learning algorithms and natural language processing, Prompt Shields effectively identifies and mitigates potential threats in user prompts and third-party data. This cutting-edge capability will support the security and integrity of your AI applications, helping to safeguard your systems against malicious attempts at manipulation or exploitation.
Prompt Shields Capabilities
AI Content Safety Prompt Shields offers numerous benefits. In addition to defending against jailbreaks, prompt injections, and document attacks, it can help to ensure that LLMs behave as designed, by blocking prompts that explicitly try to circumvent rules and policies defined by the developer. The following use cases and customer testimonials highlight the impact of these capabilities.
Customer Successes
AXA, a global leader in insurance,uses Azure OpenAI to power its Secure GPT solution. By integrating Azure’s content filtering technology and adding its own security layer, AXA prevents prompt injection attacks and helps ensure the reliability of its AI models. Secure GPT is based onAzure OpenAI in Foundry Models, taking advantage of models that have already been fine-tuned using human feedback reinforcement learning. In addition, AXA can also rely onAzure content filtering technology, to which the company added its own security layer to prevent any jailbreaking of the model using Prompt Shields, ensuring an optimal level of reliability. These layers are regularly updated to maintain advanced safeguarding.
Wrtn Technologies, a leading enterprise in Korea,relies on Azure AI Content Safetyto maintain compliance and security across its products. At its core, Wrtn’s flagship technology compiles an array of AI use cases and services localized for Korean users to integrate AI into their everyday lives. The platform fuses elements of AI-powered search, chat functionality, and customizable templates, empowering users to interact seamlessly with an “Emotional Companion” AI-infused agent. These AI agents have engaging, lifelike personalities, interacting in conversation with their creators. The vision is a highly interactive personal agent that’s unique and specific to you, your data, and your memories. Because the product is highly customizable to specific users, the built-in ability to toggle content filters and Prompt Shields is highly advantageous, allowing Wrtn to efficiently customize its security measures for different end users. This lets developers scale products while staying compliant, customizable, and responsive to users across Korea.
For IT decision makers looking to enhance the security of their AI deployments, integrating Azure’s Prompt Shields is a strategic imperative. Fortunately, enabling Prompt Shields is easy. Azure’s Prompt Shields and built-in AI security features offer an unparalleled level of protection for AI models, helping to ensure that organizations can harness the power of AI without compromising on security. Microsoft isa leader in identifying and mitigating prompt injection attacks, and uses best practices developed with decades of research, policy, product engineering, and learnings from building AI products at scale, so you can achieve your AI transformation with confidence.
By integrating these capabilities into your AI strategy, you can help safeguard your systems from prompt injection attacks and help maintain the trust and confidence of your users. Organizations across industries are usingAzure AI FoundryandMicrosoft 365 Copilotcapabilities to drive growth, increase productivity, and create value-added experiences. We’re committed to helping organizations use and build AI that istrustworthy, meaning it is secure, private, and safe. Trustworthy AI is only possible when you combine our commitments, such as ourSecure Future InitiativeandResponsible AI principles, with our product capabilities to unlock AI transformation with confidence.