To learn more about the various ways that these models can be deployed, and to see a summary of pricing and context window length information for the models, see Supported foundation models.
The foundation models in watsonx.ai support a range of use cases for both natural languages and programming languages. To see the types of tasks that these models can do, review and try the sample prompts.
allam-1-13b-instruct
Copy link to section
The allam-1-13b-instruct foundation model is a bilingual large language model for Arabic and English provided by the National Center for Artificial Intelligence and supported by the Saudi Authority for Data and Artificial Intelligence that
is fine-tuned to support conversational tasks. The ALLaM series is a collection of powerful language models designed to advance Arabic language technology. These models are initialized with Llama-2 weights and undergo training on both Arabic
and English languages.
Note:
When you inference this model from the Prompt Lab, disable AI guardrails.
Usage
Supports Q&A, summarization, classification, generation, extraction, and translation in Arabic.
allam-1-13b-instruct is based on the Allam-13b-base model, which is a foundation model that is pre-trained on a total of 3 trillion tokens in English and Arabic, including the tokens seen from its initialization. The Arabic dataset contains
500 billion tokens after cleaning and deduplication. The additional data is collected from open source collections and web crawls. The allam-1-13b-instruct foundation model is fine-tuned with a curated set of 4 million Arabic and 6 million
English prompt-and-response pairs.
Note: The maximum new tokens, which means the tokens that are generated by the foundation model per request, is limited to 8,192.
Supported natural languages
English
Supported programming languages
The codellama-34b-instruct-hf foundation model supports many programming languages, including Python, C++, Java, PHP, Typescript (Javascript), C#, Bash, and more.
Instruction tuning information
The instruction fine-tuned version was fed natural language instruction input and the expected output to guide the model to generate helpful and safe answers in natural language.
The distilled variants of the DeepSeek-R1 models based on the Llama 3.1 models are provided by DeepSeek AI. The DeepSeek-R1 models are open-sourced models with powerful reasoning capabilities. The data samples generated by the DeepSeek R1
model are used to fine tune a base Llama model.
The deepseek-r1-distill-llama-8b and deepseek-r1-distill-llama-70b models are distilled versions of the DeepSeek-R1 model based on the Llama 3.1 8B and the Llama 3.3 70B models respectively.
Usage
General use with zero- or few-shot prompts and are designed to excel in instruction-following tasks such as summarization, classification, reasoning, code tasks, as well as math.
8b and 70b: Context window length (input + output): 131,072
Note: The maximum new tokens, which means the tokens generated by the foundation model per request, is limited to 32,768.
Supported natural languages
English
Instruction tuning information
The DeepSeek-R1 models are trained by using large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step. The subsequent RL and SFT stages aim to improve reasoning patterns and align the model with
human preferences. DeepSeek-R1-Distill models are fine-tuned based on open-source models, using samples generated by DeepSeek-R1.
The elyza-japanese-llama-2-7b-instruct model is provided by ELYZA, Inc on Hugging Face. The elyza-japanese-llama-2-7b-instruct foundation model is a version of the Llama 2 model from Meta that is trained to understand
and generate Japanese text. The model is fine-tuned for solving various tasks that follow user instructions and for participating in a dialog.
Usage
General use with zero- or few-shot prompts. Works well for classification and extraction in Japanese and for translation between English and Japanese. Performs best when prompted in
Japanese.
For Japanese language training, Japanese text from many sources were used, including Wikipedia and the Open Super-large Crawled ALMAnaCH coRpus (a multilingual corpus that is generated by classifying and filtering language in the Common
Crawl corpus). The model was fine-tuned on a dataset that was created by ELYZA. The ELYZA Tasks 100 dataset contains 100 diverse and complex tasks that were created manually and evaluated by humans. The ELYZA Tasks 100 dataset
is publicly available from HuggingFace.
The EuroLLM series of models is developed by the Unified Transcription and Translation for Extended Reality (UTTER) Project and the European Union. The EuroLLM Instruct models are open-source models specialized in understanding and generating
text across all the 24 official European Union (EU) languages, as well as 11 commercially and strategically important international languages.
Usage
Suited for multingual language tasks like general instructon-following and language translation.
The models are trained on 4 trillion tokens across the supported natural languages from web data, parallel data, Wikipedia, Arxiv, multiple books, and Apollo datasets.
The flan-t5-xl-3b model is provided by Google on Hugging Face. The model is based on the pretrained text-to-text transfer transformer (T5) model and uses instruction fine-tuning methods to achieve better zero- and
few-shot performance. The model is also fine-tuned with chain-of-thought data to improve its ability to perform reasoning tasks.
Note:
This foundation model can be tuned by using the Tuning Studio.
The model was fine-tuned on tasks that involve multiple-step reasoning from chain-of-thought data in addition to traditional natural language processing tasks. Details about the training datasets used are published.
The flan-t5-xxl-11b model is provided by Google on Hugging Face. This model is based on the pretrained text-to-text transfer transformer (T5) model and uses instruction fine-tuning methods to achieve better zero- and few-shot performance. The model is also fine-tuned with chain-of-thought data to improve its ability to perform reasoning tasks.
The model was fine-tuned on tasks that involve multiple-step reasoning from chain-of-thought data in addition to traditional natural language processing tasks. Details about the training datasets used are published.
The flan-ul2-20b model is provided by Google on Hugging Face. This model was trained by using the Unifying Language Learning Paradigms (UL2). The model is optimized for language generation, language understanding, text classification, question
answering, common sense reasoning, long text reasoning, structured-knowledge grounding, and information retrieval, in-context learning, zero-shot prompting, and one-shot prompting.
The flan-ul2-20b model is pretrained on the colossal, cleaned version of Common Crawl's web crawl corpus. The model is fine-tuned with multiple pretraining objectives to optimize it for various natural language processing tasks. Details
about the training datasets used are published.
Jais-13b-chat is based on the Jais-13b model, which is a foundation model that is trained on 116 billion Arabic tokens and 279 billion English tokens. Jais-13b-chat is fine tuned with a curated set of 4 million Arabic and 6 million English
prompt-and-response pairs.
Tech preview of the Llama 4 collection of foundation models that are provided by Meta. The llama-4-maverick-17b-128e-instruct-fp8 and llama-4-scout-17b-16e-instruct models are multimodal models that use a mixture-of-experts
(MoE) architecture for optimized, best-in-class performance in text and image understanding.
The Llama 4 Maverick model is a 17 billion active parameter multimodal model with 128 experts. The Llama 4 Scout model is a 17 billion active parameter multimodal model with 16 experts.
Usage
Generates multilingual dialog output like a chatbot, uses a model-specific prompt format, optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.
Size
17 billion parameters
API pricing tier
These models are available as preview models with no charge.
Llama 4 was pre-trained on a broader collection of 200 languages. The Llama 4 Scout model was pre-trained on approximately 40 trillion tokens and the Llama 4 Maverick model was pre-trained on approximately 22 trillion tokens of multimodal
data from publicly available and licensed information from Meta.
The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model (text in/text out) with 70 billion parameters.
The llama-3-3-70b-instruct is a revision of the popular Llama 3.1 70B Instruct foundation model. The Llama 3.3 foundation model is better at coding, step-by-step reasoning, and tool-calling. Despite its smaller size, the Llama 3.3 model's
performance is similar to that of the Llama 3.1 405b model, making it a great choice for developers.
Usage
Generates multilingual dialog output like a chatbot. Uses a model-specific prompt format.
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai
Instruction tuning information
Llama 3.3 was pretrained on 15 trillion tokens of data from publicly available sources. The fine tuning data includes publicly available instruction datasets, as well as over 25 million synthetically generated examples.
The Llama 3.2 collection of foundation models are provided by Meta. The llama-3-2-1b-instruct and llama-3-2-3b-instruct models are the smallest Llama 3.2 models that fit onto a mobile device. The models are lightweight,
text-only models that can be used to build highly personalized, on-device agents.
For example, you can ask the models to summarize the last ten messages you received, or to summarize your schedule for the next month.
Usage
Generate dialog output like a chatbot. Use a model-specific prompt format. Their small size and modest compute resource and memory requirements enable the Llama 3.2 Instruct models to be run locally on most hardware, including on mobile
and other edge devices.
The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 8,192.
Supported natural languages
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai
Instruction tuning information
Pretrained on up to 9 trillion tokens of data from publicly available sources. Logits from the Llama 3.1 8B and 70B models were incorporated into the pretraining stage of the model development, where outputs (logits) from these larger
models were used as token-level targets. In post-training, aligned the pre-trained model by using Supervised Fine-Tuning (SFT), Rejection Sampling (RS), and Direct Preference Optimization (DPO).
The Meta Llama 3.2 collection of foundation models are provided by Meta. The llama-3-2-11b-vision-instruct and llama-3-2-90b-vision-instruct models are built for image-in, text-out use
cases such as document-level understanding, interpretation of charts and graphs, and captioning of images.
Usage
Generates dialog output like a chatbot and can perform computer vision tasks including classification, object detection and identification, image-to-text transcription (including handwriting), contextual Q&A, data extraction and processing,
image comparison and personal visual assistance. Uses a model-specific prompt format.
The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 8,192. The tokens that are counted for an image that you submit to the model are not included in the context window length.
Supported natural languages
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai with text-only inputs. English only when an image is included with the input.
Instruction tuning information
Llama 3.2 Vision models use image-reasoning adaptor weights that are trained separately from the core large language model weights. This separation preserves the general knowledge of the model and makes the
model more efficient both at pretraining time and run time. The Llama 3.2 Vision models were pretrained on 6 billion image-and-text pairs, which required far fewer compute resources than were needed to pretrain
the Llama 3.1 70B foundation model alone. Llama 3.2 models also run efficiently because they can tap more compute resources for image reasoning only when the input requires it.
The Meta Llama 3.2 collection of foundation models are provided by Meta. The llama-guard-3-11b-vision is a multimodal evolution of the text-only Llama-Guard-3 model. The model can be
used to classify image and text content in user inputs (prompt classification) as safe or unsafe.
Usage
Use the model to check the safety of the image and text in an image-to-text prompt.
The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 8,192. The tokens that are counted for an image that you submit to the model are not included in the context window length.
Supported natural languages
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai with text-only inputs. English only when an image is included with the input.
Instruction tuning information
Pretrained model that is fine-tuned for content safety classification. For more information about the types of content that are classified as unsafe, see the model card.
The Meta Llama 3.1 collection of foundation models are provided by Meta. The Llama 3.1 base foundation models, llama-3-1-8b and llama-3-1-70b are multilingual models that supports tool
use, and have overall stronger reasoning capabilities.
Usage
Use for long-form text summarization and with multilingual conversational agents or coding assistants.
The Meta Llama 3.1 collection of foundation models are provided by Meta. The Llama 3.1 foundation models are pretrained and instruction tuned text-only generative models that are optimized
for multilingual dialogue use cases. The models use supervised fine-tuning and reinforcement learning with human feedback to align with human preferences for helpfulness and safety.
The llama-3-405b-instruct model is Meta's largest open-sourced foundation model to date. This foundation model can also be used as a synthetic data generator, post-training data ranking judge, or model teacher/supervisor
that can improve specialized capabilities in more inference-friendly, derivative models.
Usage
Generates dialog output like a chatbot. Uses a model-specific prompt format.
Although the model supports a context window length of 131,072, the window is limited to 16,384 to reduce the time it takes for the model to generate a response.
The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 4,096.
Supported natural languages
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai
Instruction tuning information
Llama 3.1 was pretrained on 15 trillion tokens of data from publicly available sources. The fine tuning data includes publicly available instruction datasets, as well as over 25 million synthetically generated examples.
The Meta Llama 3 family of foundation models are accessible, open large language models that are built with Meta Llama 3 and provided by Meta on Hugging Face.
The Llama 3 foundation models are instruction fine-tuned language models that can support various use cases.
Note: The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 4,096.
Supported natural languages
English
Instruction tuning information
Llama 3 features improvements in post-training procedures that reduce false refusal rates, improve alignment, and increase diversity in the foundation model output. The result is better reasoning, code generation, and instruction-following
capabilities. Llama 3 has more training tokens (15T) that result in better language comprehension.
The Llama 2 Chat models are provided by Meta on Hugging Face. The fine-tuned models are useful for chat generation. The models are pretrained with publicly available online data and fine-tuned using reinforcement
learning from human feedback.
You can choose to use the 13 billion parameter or 70 billion parameter version of the model.
Usage
Generates dialog output like a chatbot. Uses a model-specific prompt format.
Llama 2 was pretrained on 2 trillion tokens of data from publicly available sources. The fine-tuning data includes publicly available instruction datasets and more than one million new examples that were annotated by humans.
Mistral Large 2 is a family of large language models developed by Mistral AI. The mistral-large foundation model is fluent in and understands the grammar and cultural context of English, French, Spanish, German,
and Italian. The foundation model can also understand dozens of other languages. The model has a large context window, which means you can add large documents as contextual information in prompts that you submit for retrieval-augmented generation
(RAG) use cases. The mistral-large foundation model is effective at programmatic tasks, such as generating, reviewing, and commenting on code, function calling, and can generate results in JSON format.
For more getting started information, see the watsonx.ai page on the Mistral AI website.
Usage
Suitable for complex multilingual reasoning tasks, including text understanding, transformation, and code generation. Due to the model's large context window, use the max tokens parameter to specify a token limit when prompting the model.
API pricing tier
Pricing for inferencing the provided Mistral Large model is not assigned by a multiplier. The following special pricing tiers are used:
Input tier: Mistral Large Input
Output tier: Mistral Large
For pricing details, see Table 3. For pricing details for deploying this model on demand, see Table 5.
Attention: This foundation model has an additional access fee that is applied per hour of use.
The mistral-large-instruct-2411 foundation model from Mistral AI and belongs to the Mistral Large 2 family of models. The model specializes in reasoning, knowledge, and coding. The model extends the capabilities
of the Mistral-Large-Instruct-2407 foundation model to include better handling of long prompt contexts, system prompt instructions, and function calling requests.
Usage
The mistral-large-instruct-2411 foundation model is multilingual, proficient in coding, agent-centric, and adheres to system prompts to aid in retrieval-augmented generation tasks and other use cases where prompts with large context need
to be handled.
Multiple languages and is particularly strong in English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.
Supported programming languages
The mistral-large-instruct-2411 foundation model has been trained on over 80 programming languages including Python, Java, C, C++, JavaScript, Bash, Swift, and Fortran.
Instruction tuning information
The mistral-large-instruct-2411 foundation model extends the Mistral-Large-Instruct-2407 foundation model from Mistral AI. Training enhanced the reasoning capabilities of the model. Training also focused on
reducing hallucinations by fine tuning the model to be more cautious and discerning in its responses and to acknowledge when it cannot find solutions or does not have sufficient information to provide a confident answer.
License
For terms of use, including information about contractual protections related to capped indemnification, see Terms of use.
The mistral-nemo-instruct-2407 foundation model from Mistral AI was built in collaboration with NVIDIA. Mistral NeMo performs exceptionally well in reasoning,
world knowledge, and coding accuracy, especially for a model of its size.
Usage
The Mistral NeMo model is multilingual and is trained on function calling.
Mistral Small 3 is a cost-efficient, fast, and reliable foundation model developed by Mistral AI. The mistral-small-24b-instruct-2501 model is instruction fine-tuned and performs well in tasks that require some
reasoning ability, such as data extraction, summarizing a document, or writing descriptions. Built to support agentic application, with adherence to system prompts and function calling with JSON output generation.
For more getting started information, see the watsonx.ai page on the Mistral AI website.
Usage
Suitable for conversational agents and function calling.
The maximum new tokens, which means the tokens generated by the foundation model per request, is limited to 16,384.
Supported natural languages
English, French, German, Italian, Spanish, Chinese, Japanese, Korean, Portuguese, Dutch, Polish, and dozens of other languages.
Supported programming languages
The mistral-small-24b-instruct-2501 model has been trained on over 80 programming languages including Python, Java, C, C++, JavaScript, Bash, Swift, and Fortran.
Instruction tuning information
The mistral-small-24b-instruct-2501 foundation model is pre-trained on diverse datasets like text, codebases, and mathematical data from various domains.
The mixtral-8x7b-base foundation model is provided by Mistral AI. The mixtral-8x7b-base foundation model is a generative sparse mixture-of-experts network that groups the model parameters, and then for each token
chooses a subset of groups (referred to as experts) to process the token. As a result, each token has access to 47 billion parameters, but only uses 13 billion active parameters for inferencing, which reduces costs and latency.
Usage
Suitable for many tasks, including classification, summarization, generation, code creation and conversion, and language translation.
The mixtral-8x7b-instruct-v01 foundation model is provided by Mistral AI. The mixtral-8x7b-instruct-v01 foundation model is a pretrained generative sparse mixture-of-experts network that groups the model parameters,
and then for each token chooses a subset of groups (referred to as experts) to process the token. As a result, each token has access to 47 billion parameters, but only uses 13 billion active parameters for inferencing, which reduces
costs and latency.
Usage
Suitable for many tasks, including classification, summarization, generation, code creation and conversion, and language translation. Due to the model's unusually large context window, use the max tokens parameter to specify a token limit
when prompting the model.
The mt0-xxl-13b model is provided by BigScience on Hugging Face. The model is optimized to support language generation and translation tasks with English, languages other than English, and multilingual prompts.
Usage: General use with zero- or few-shot prompts. For translation tasks, include a period to indicate the end of the text you want translated or the model might continue
the sentence rather than translate it.
Pixtral 12B is a multimodal model developed by Mistral AI. The pixtral-12b foundation model is trained to understand both natural images and documents and is able to ingest images at their natural resolution and
aspect ratio, providing flexibility on the number of tokens used to process an image. The foundation model supports multiple images in its long context window. The model is effective in image-in, text-out multimodal tasks and excels at instruction
following.
Usage
Chart and figure understanding, document question answering, multimodal reasoning, and instruction following.
The maximum new tokens, which means the tokens generated by the foundation models per request, is limited to 8,192.
Supported natural languages
English
Instruction tuning information
The pixtral-12b model is trained with interleaved image and text data and is based on the Mistral Nemo model with a 400 million parameter vision encoder trained from scratch.
Any deprecated foundation models are highlighted with a deprecated warning icon . For more information about deprecation, including foundation
model withdrawal details, see Foundation model lifecycle.