You can analyze model transactions to understand how your model predicts outcomes. For example, you can review the predicted biased outcomes in payload transactions, or analyze a drop in accuracy, to get a better understanding of how your deployment
is performing.
The following different methods are provided to help you analyze your model transactions:
Reviewing payload transactions
Copy link to section
When you review results from fairness evaluations, you can select View payload transactions to view a list of all of the model transactions that are processed to predict
any biased outcomes for each monitored attribute.
Reviewing drift transactions
Copy link to section
When you review results from drift evaluations, you can click the Number of transactions link on a transactions summary tile to view the transactions that are responsible
for drift. The Transactions page specifies reasons for the drop in accuracy or drop in data consistency and provides a recommendation to help fix the drift of your model.
Reviewing generative AI quality transactions
Copy link to section
When you review results from generative AI quality evaluations, you can select an event on the data visualizations for metric scores to view a table of transactions. You can select a bar on the score distribution chart to filter the range of
transactions that are displayed in the table.
To review results for RAG tasks, you can select events on the data visualizations for answer and retrieval quality metrics to display a venn diagram that helps you analyze transactions for threshold violations. You can select areas in the venn
diagram to view a table of transactions that corresponds to the metric results that you choose to review for threshold violations. You can select the Analyze action in the transactions table to view source attributions for
each transaction. To gain insights about why threshold violations are detected in your evaluation results, you can also select See root cause analysis and select See recommended actions to help you understand
how to improve your model performance.
Reviewing performance
Copy link to section
When you review evaluation results, you can review performance to understand how your model processes scoring requests. Performance evaluations are enabled by default when you add deployments. To evaluate
performance, the total number of transaction records and the average number of transaction records per minute that your model processes is calculated.
Reviewing analytics
Copy link to section
When you review evaluation results, you can use the Predictions by confidence chart to analyze the transactions that are sent to your deployment in the date range that you select by reviewing
prediction classes and confidence distribution in each class. You can also use the Chart builder to create custom visualizations to help you better understand model predictions and inputs at run time.
The chart builder displays the output of the model’s prediction against the features or data ranges that you specify. For example, you can use the chart builder to see the split in predicted classes or confidence scores for different ranges
of an attribute. You can also analyze the transactions that are sent to your deployment during a data range by creating a custom chart and choosing between featues, prediction classes, and confidence.
When you review any transactions on the Transactions page, you can choose to view explanations that provide details about how the predictions for each transaction are determined. For more information, see Explaining model transactions.
Note:
If your data includes a feature name containing a period (.), rendering a chart will fail with this error: Failed to get data
About cookies on this siteOur websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising.For more information, please review your cookie preferences options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.