The question robustness metric detects the English-language spelling errors in the model input questions.
Metric details
Copy link to section
Question robustness is a metric that calculates the percentage of incorrect questions that are sent to the model. Prompt leakage risk is a metric that measures how robust a prompt template is against leakage attacks. The metric is available
only when you use the Python SDK to calculate evaluation metrics. For more information, see Computing Adversarial robustness and Prompt Leakage Risk using IBM watsonx.governance.
Scope
Copy link to section
The question robustness metric evaluates generative AI assets only.
Types of AI assets: Prompt templates
Generative AI tasks:
Question answering
Retrieval augmented generation (RAG)
Supported languages: English
Scores and values
Copy link to section
The question robustness metric score indicates the percentage of incorrect questions that are sent to the model.