0 / 0
Configuring model evaluations
Last updated: Oct 25, 2024
Configuring model evaluations

You can configure evaluations to generate insights about your model performance.

You can configure the following types of evaluations:

  • Quality
    Evaluates how well your model predicts correct outcomes that match labeled test data.
  • Fairness
    Evaluates whether your model produces biased outcomes that provide favorable results for one group over another.
  • Drift
    Evaluates how your model changes in accuracy and data consistency by comparing recent transactions to your training data.
  • Drift v2
    Evaluates changes in your model output, the accuracy of your predictions, and the distribution of your input data.
  • Model health
    Evaluates how efficiently your model deployment processes your transactions.
  • Generative AI quality
    Measures how well your foundation model performs tasks

If you're evaluating traditional machine learning models, you can also create custom evaluations and metrics to generate a greater variety of insights about your model performance.

Each evaluation generates metrics that you can analyze to gain insights about your model performance.

When you configure evaluations, you can choose to run evaluations continuously on the following default scheduled intervals:

Evaluation Online subscription default schedule Batch subscription default schedule
Quality 1 hour 1 week
Fairness 1 hour 1 week
Drift 3 hours 1 week
Drift v2 1 day NA
Explainability 1 week 1 week
Model health 1 hour NA
Generative AI Quality 1 hour NA

Parent topic: Evaluating AI models with Watson OpenScale

Generative AI search and answer
These answers are generated by a large language model in watsonx.ai based on content from the product documentation. Learn more