Prompt Shields (Azure)
This evaluator leverages Azure's Prompt Shields API to detect and prevent adversarial inputs targeting user prompts and documents. It's specially designed to enhance the safety of interactions with language models by analyzing and flagging potential attacks.
Configuration
No specific configuration parameters are required for this evaluator as it operates with default settings designed to optimize performance and security.
Metrics
This evaluator reports the following metrics to provide insights into detected adversarial activities:
Name | Key | Description | Min | Max |
---|---|---|---|---|
User Prompt Attack | user_prompt_attack | Indicates whether the input text contains a user prompt attack | 0 | 1 |
Document Attack | document_attack | Indicates whether any external documents contain attacks | 0 | 1 |
Additional Information
Please refer to the official Azure documentation for a more comprehensive understanding of how Azure Prompt Shields works and how to implement it within your systems.
Last updated