Prompt Shields (Azure)

This evaluator leverages Azure's Prompt Shields API to detect and prevent adversarial inputs targeting user prompts and documents. It's specially designed to enhance the safety of interactions with language models by analyzing and flagging potential attacks.

Configuration

No specific configuration parameters are required for this evaluator as it operates with default settings designed to optimize performance and security.

Metrics

This evaluator reports the following metrics to provide insights into detected adversarial activities:

NameKeyDescriptionMinMax

User Prompt Attack

user_prompt_attack

Indicates whether the input text contains a user prompt attack

0

1

Document Attack

document_attack

Indicates whether any external documents contain attacks

0

1

Additional Information

Please refer to the official Azure documentation for a more comprehensive understanding of how Azure Prompt Shields works and how to implement it within your systems.

Last updated