Choose the hardware configuration for your watsonx.ai Runtime asset when you train the asset or when you deploy it.
Hardware configurations available for training and deploying assets
Capacity type
Capacity units per hour
Extra small: 1x4 = 1 vCPU and 4 GB RAM
0.5
Small: 2x8 = 2 vCPU and 8 GB RAM
1
Medium: 4x16 = 4 vCPU and 16 GB RAM
2
Large: 8x32 = 8 vCPU and 32 GB RAM
4
Extra large: 16x64 = 16 vCPU and 64 GB RAM
8
Compute usage for watsonx.ai Runtime assets
Copy link to section
Deployments and scoring consume compute resources as capacity unit hours (CUH) from the watsonx.ai Runtime service.
To check the total monthly CUH consumption for your watsonx.ai Runtime services, from the navigation menu, select Administration -> Environment runtimes.
Additionally, you can monitor the monthly resource usage in each specific deployment space. To do that, from your deployment space, go to the Manage tab and then select Resource usage. The summary shows CUHs
used by deployment type: separately for AutoAI deployments, Federated Learning deployments, batch deployments, and online deployments.
Compute usage details
Copy link to section
The rate of consumed CUHs is determined by the computing requirements of your deployments. It is based on such variables as:
type of deployment
type of framework
complexity of scoring Scaling a deployment to support more concurrent users and requests also increases CUH consumption. As many variables affect resource consumption for a deployment, it is recommended that you run tests on your models
and deployments to analyze CUH consumption.
The way that online deployments consume capacity units is based on framework. For some frameworks, CUHs are charged for the number of hours that the deployment asset is active in a deployment space. For example, SPSS models in online deployment
mode that run for 24 hours a day, seven days a week, consume CUHs and are charged for that period. An active online deployment has no idle time. For other frameworks, CUHs are charged according to scoring duration. Refer to the CUH consumption
table for details on how CUH usage is calculated.
Compute time is calculated to the millisecond, with a 1-minute minimum for each distinct operation. For example:
A training run that takes 12 seconds is billed as 1 minute
A training run that takes 83.555 seconds is billed exactly as calculated
CUH consumption by deployment and framework type
Copy link to section
CUH consumption is calculated by using these formulas:
Deployment type
Framework
CUH calculation
Online
AutoAI, AI function, SPSS, Scikit-Learn custom libraries, Tensorflow, RShiny
Deployment active duration * Number of nodes * CUH rate for capacity type framework
Online
Spark, PMML, Scikit-Learn, Pytorch, XGBoost
Score duration in seconds * Number of nodes * CUH rate for capacity type framework
Batch
all frameworks
Job duration in seconds * Number of nodes * CUH rate for capacity type framework
About cookies on this siteOur websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising.For more information, please review your cookie preferences options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.