Requirements and usage notes for custom foundation models
Copy link to section
Deployable custom models must meet these requirements:
Uploading and using your own custom model is available only in the Standard plan for watsonx.ai.
The model must be compatible with the Text Generation Inference (TGI) standard and be built with a supported model architecture type.
The file list for the model must contain a config.json file.
The model must be in a safetensors format with the supported transformers library and must include a tokenizer.json file.
Important:
You must make sure that your custom foundation model is saved with the supported transformers library. If the model.safetensors file for your custom foundation model uses an unsupported data format in the metadata header,
your deployment might fail. For more information, see Troubleshooting watsonx.ai Runtime.
Make sure that the project or space where you want to deploy your custom foundation model has an associated watsonx.ai Runtime instance. Open the Manage tab in your project or space to check that.
Supported model architectures
Copy link to section
The following table lists the model architectures that you can deploy as custom models for inferencing with watsonx.ai. The model architectures are listed together with information about their supported quantization methods, parallel tensors,
deployment configuration sizes, and software specifications.
Note:
Two software specifications are available for your deployments: watsonx-cfm-caikit-1.0, watsonx-cfm-caikit-1.1. The watsonx-cfm-caikit-1.1 specification is better in terms of performance, but it's not
available with every model architecture.
Supported model architectures, quantization methods, parallel tensors, deployment configuration sizes, and software specifications
IBM does not support deployment failures as a result of deploying foundation models with unsupported architectures.
Deployments of llama 3.1 models might fail. To address this issue, see steps that are listed in Troubleshooting.
It is not possible to deploy codegen, mt5, and t5 type models with the watsonx-cfm-caikit-1.1 software specification.
If your model does not support parallel tensors, the only configuration that you can use is Small. If your model was trained with more parameters than the Small configuration supports it will fail. This means
that you won't be able to deploy some of your custom models. For more information on limitations, see Resource utilization guidelines.
Collecting the prerequisite details for a custom foundation model
Copy link to section
Check for the existence of the file config.json in the foundation model content folder. Deployment service will mandate for existence of the file config.json in the foundation model content folder after it is uploaded
to the cloud storage.
Open the config.json file to confirm that the foundation model uses a supported architecture.
View the list of files for the foundation model to check for the tokenizer.json file and that the model content is in .safetensors format.
Important:
You must make sure that your custom foundation model is saved with the supported transformers library. If the model.safetensors file for your custom foundation model uses an unsupported data format in the metadata header,
your deployment might fail. For more information, see Troubleshooting watsonx.ai Runtime.
See an example:
For the falcon-40b model stored on Hugging Face, click Files and versions to view the file structure and check for config.json:
The example model uses a version of the supported falcon architecture.
This example model contains the tokenizer.json file and is in the .safetensors format:
If the model does not meet these requirements, you cannot create a model asset and deploy your model.
Resource utilization guidelines
Copy link to section
Three configurations are available to support your custom foundation model: Small, Medium, and Large. To determine the most suitable configuration for your custom foundation model, see the following guidelines:
Assign the Small configuration to any double-byte precision model under 26B parameters, subject to testing and validation.
Assign the Medium configuration to any double-byte precision model between 27B and 53B parameters, subject to testing and validation.
Assign the Large configuration to any double byte precision model between 54B and 106B parameters, subject to testing and validation.
Tip:
If the selected configuration fails during the testing and validation phase, consider exploring the next higher configuration available. For example, try the Medium configuration if the Small configuration fails.
Currently the Large configuration is the highest available configuration.
Limitation and restrictions for custom foundation models
Copy link to section
Note these limits on how you can deploy and use custom foundation models with watsonx.ai.
Limitations for deploying custom foundation models
Copy link to section
Due to high demand for custom foundation model deployments and limited resources to accommodate it, watsonx.ai has a deployment limit of either four small models, two medium models, or one large model per IBM Cloud account. If you attempt
to import a custom foundation model beyond these limits, you will be notified and asked to share your feedback through a survey. This will help us understand your needs and plan for future capacity upgrades.
Important: Any requested limit increases are not guaranteed and are subject to available capacity.
Restrictions for using custom foundation model deployments
Copy link to section
Note these restrictions for using custom foundation models after they are deployed with watsonx.ai:
You cannot tune a custom foundation model.
You cannot use watsonx.governance to evaluate or track a prompt template for a custom foundation model.
You can prompt a custom foundation model but cannot save a prompt template for a custom model.
Help us improve this experience
Copy link to section
If you want to share your feedback now, click this link. Your feedback is essential in helping us plan for future capacity upgrades and improve the overall custom foundation
model deployment experience. Thank you for your cooperation!
About cookies on this siteOur websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising.For more information, please review your cookie preferences options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.