You can work with third-party and IBM foundation models in IBM watsonx.ai. You can use foundation models that are provided by IBM and are ready to use immediately, or deploy foundation models on-demand to use exclusively for your organization.
The list of all the foundation models you can use in watsonx.ai is as follows:
- granite-3-2-8b-instruct
- granite-3-2-8b-instruct-preview-rc
- granite-3-2b-instruct
- granite-3-8b-instruct
- granite-3-8b-base
- granite-7b-lab
- granite-8b-japanese
- granite-13b-chat-v2
- granite-13b-instruct-v2
- granite-20b-multilingual
- granite-3b-code-instruct
- granite-8b-code-instruct
- granite-20b-code-instruct
- granite-20b-code-base-schema-linking
- granite-20b-code-base-sql-gen
- granite-34b-code-instruct
- granite-guardian-3-2b
- granite-guardian-3-8b
- granite-ttm-512-96-r2
- granite-ttm-1024-96-r2
- granite-ttm-1536-96-r2
- granite-vision-3-2-2b
- allam-1-13b-instruct
- codellama-34b-instruct-hf
- deepseek-r1-distill-llama-8b
- deepseek-r1-distill-llama-70b
- elyza-japanese-llama-2-7b-instruct
- flan-t5-xl-3b
- flan-t5-xxl-11b
- flan-ul2-20b
- jais-13b-chat
- llama-3-3-70b-instruct
- llama-3-2-1b-instruct
- llama-3-2-3b-instruct
- llama-3-2-11b-vision-instruct
- llama-3-2-90b-vision-instruct
- llama-guard-3-11b-vision
- llama-3-1-8b
- llama-3-1-8b-instruct
- llama-3-1-70b-instruct
- llama-3-405b-instruct
- llama-3-8b-instruct
- llama-3-70b-instruct
- llama-2-13b-chat
- llama-2-70b-chat
- mistral-large
- mistral-large-instruct-2407
- mistral-large-instruct-2411
- mistral-nemo-instruct-2407
- mistral-small-24b-instruct-2501
- mixtral-8x7b-base
- mixtral-8x7b-instruct-v01
- mt0-xxl-13b
- pixtral-12b
How to choose a model
To review factors that can help you to choose a model, such as supported tasks and languages, see Choosing a model and Foundation model benchmarks.
Foundation models by deployment method
Depending on the deployment method, you can use foundation models on multitenant hardware directly or deploy models on dedicated hardware for use by your organization. To learn more about the various ways you can use to deploy models, see Foundation model deployment methods.
Provided foundation models that are ready to use
A collection of open source and IBM foundation models are deployed in IBM watsonx.ai. You can prompt these foundation models in the Prompt Lab or programmatically.
IBM foundation models
The following table lists the supported IBM foundation models that IBM provides for inferencing.
Use is measured in Resource Units (RU); each unit is equal to 1,000 tokens from the input and output of foundation model inferencing. For details on how model pricing is calculated and monitored, see Billing details for generative AI assets.
You can also access some IBM foundation models from third-party repositories, such as Hugging Face. IBM foundation models that you obtain from a third-party repository are not indemnified by IBM. Only IBM foundation models that you access from watsonx.ai are indemnified by IBM. For more information about contractual protections related to IBM indemnification, see the IBM Client Relationship Agreement and IBM watsonx.ai service description.
Model name | API model_id | Input price (USD/1,000 tokens) |
Output price (USD/1,000 tokens) |
Context window (input + output tokens) |
More information |
---|---|---|---|---|---|
granite-13b-instruct-v2 | ibm/granite-13b-instruct-v2 |
$0.0006 | $0.0006 | 8,192 | • Model card • Website • Research paper Note: This foundation model can be prompt tuned.
|
granite-8b-japanese | ibm/granite-8b-japanese |
$0.0006 | $0.0006 | 4,096 | • Model card • Website • Research paper |
granite-20b-multilingual | ibm/granite-20b-multilingual |
$0.0006 | $0.0006 | 8,192 | • Model card • Website • Research paper |
granite-3-2b-instruct | ibm/granite-3-2b-instruct |
$0.0001 | $0.0001 | 131,072 | • Model card • Website • Research paper |
granite-3-8b-instruct | ibm/granite-3-8b-instruct |
$0.0002 | $0.0002 | 131,072 | • Model card • Website • Research paper |
granite-3-2-8b-instruct | ibm/granite-3-2-8b-instruct |
$0.0002 | $0.0002 | 131,072 | • Model card • Website • Research paper |
granite-3-2-8b-instruct-preview-rc |
ibm/granite-3-2-8b-instruct-preview-rc |
No cost during tech preview | No cost during tech preview | 131,072 | • Model card • Website • Research paper |
granite-guardian-3-2b | ibm/granite-guardian-3-2b |
$0.0001 | $0.0001 | 131,072 | • Model card • Website |
granite-guardian-3-8b | ibm/granite-guardian-3-8b |
$0.0002 | $0.0002 | 131,072 | • Model card • Website |
granite-3b-code-instruct | ibm/granite-3b-code-instruct |
$0.0006 | $0.0006 | 128,000 | • Model card • Website • Research paper |
granite-8b-code-instruct | ibm/granite-8b-code-instruct |
$0.0006 | $0.0006 | 128,000 | • Model card • Website • Research paper |
granite-20b-code-instruct | ibm/granite-20b-code-instruct |
$0.0006 | $0.0006 | 8,192 | • Model card • Website • Research paper |
granite-34b-code-instruct | ibm/granite-34b-code-instruct |
$0.0006 | $0.0006 | 8,192 | • Model card • Website • Research paper |
granite-vision-3-2-2b | ibm/granite-vision-3-2-2b |
$0.0001 | $0.0001 | 131,072 | • Model card • Website • Research paper |
Model name | API model_id | Input price (USD/1,000 data points) |
Output price (USD/1,000 data points) |
Context length Min data points |
More information |
---|---|---|---|---|---|
granite-ttm-512-96-r2 | ibm/granite-ttm-512-96-r2 |
$0.00013 | $0.00038 | 512 | • Model card • Website • Research paper |
granite-ttm-1024-96-r2 | ibm/granite-ttm-1024-96-r2 |
$0.00013 | $0.00038 | 1,024 | • Model card • Website • Research paper |
granite-ttm-1536-96-r2 | ibm/granite-ttm-1536-96-r2 |
$0.00013 | $0.00038 | 1,536 | • Model card • Website • Research paper |
Third-party foundation models
The following table lists the supported third-party foundation models that are provided with watsonx.ai.
Use is measured in Resource Units (RU); each unit is equal to 1,000 tokens from the input and output of foundation model inferencing. For details on how model pricing is calculated and monitored, see Billing details for generative AI assets.
Model name | API model_id | Provider | Input price (USD/1,000 tokens) |
Output price (USD/1,000 tokens) |
Context window (input + output tokens) |
More information |
---|---|---|---|---|---|---|
allam-1-13b-instruct | sdaia/allam-1-13b-instruct |
National Center for Artificial Intelligence and Saudi Authority for Data and Artificial Intelligence | $0.0018 | $0.0018 | 4,096 | • Model card |
codellama-34b-instruct | codellama/codellama-34b-instruct-hf |
Code Llama | $0.0018 | $0.0018 | 16,384 | • Model card • Meta AI Blog |
elyza-japanese-llama-2-7b-instruct | elyza/elyza-japanese-llama-2-7b-instruct |
ELYZA, Inc | $0.0018 | $0.0018 | 4,096 | • Model card • Blog on note.com |
flan-t5-xl-3b | google/flan-t5-xl |
$0.0006 | $0.0006 | 4,096 | • Model card • Research paper Note: This foundation model can be prompt tuned.
|
|
flan-t5-xxl-11b | google/flan-t5-xxl |
$0.0018 | $0.0018 | 4,096 | • Model card • Research paper |
|
flan-ul2-20b | google/flan-ul2 |
$0.0050 | $0.0050 | 4,096 | • Model card • UL2 research paper • Flan research paper |
|
jais-13b-chat | core42/jais-13b-chat |
Inception, Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), and Cerebras Systems | $0.0018 | $0.0018 | 2,048 | • Model card • Research paper |
llama-3-3-70b-instruct | meta-llama/llama-3-3-70b-instruct |
Meta | $0.00071 | $0.00071 | 131,072 | • Model card • Meta AI blog |
llama-3-2-1b-instruct | meta-llama/llama-3-2-1b-instruct |
Meta | $0.0001 | $0.0001 | 131,072 | • Model card • Meta AI blog • Research paper |
llama-3-2-3b-instruct | meta-llama/llama-3-2-3b-instruct |
Meta | $0.00015 | $0.00015 | 131,072 | • Model card • Meta AI blog • Research paper |
llama-3-2-11b-vision-instruct | meta-llama/llama-3-2-11b-vision-instruct |
Meta | $0.00035 | $0.00035 | 131,072 | • Model card • Meta AI blog • Research paper |
llama-3-2-90b-vision-instruct | meta-llama/llama-3-2-90b-vision-instruct |
Meta | $0.0020 | $0.0020 | 131,072 | • Model card • Meta AI blog • Research paper |
llama-guard-3-11b-vision | meta-llama/llama-guard-3-11b-vision |
Meta | $0.00035 | $0.00035 | 131,072 | • Model card • Meta AI blog • Research paper |
llama-3-1-8b-instruct | meta-llama/llama-3-1-8b-instruct |
Meta | $0.0006 | $0.0006 | 131,072 | • Model card • Meta AI blog |
llama-3-1-70b-instruct | meta-llama/llama-3-1-70b-instruct |
Meta | $0.0018 | $0.0018 | 131,072 | • Model card • Meta AI blog |
llama-3-405b-instruct | meta-llama/llama-3-405b-instruct |
Meta | $0.0050 | $0.016 | 16,384 | • Model card • Meta AI blog |
llama-3-70b-instruct | meta-llama/llama-3-70b-instruct |
Meta | $0.0018 | $0.0018 | 8,192 | • Model card • Meta AI blog |
llama-2-13b-chat | meta-llama/llama-2-13b-chat |
Meta | $0.0006 | $0.0006 | 4,096 | • Model card • Research paper |
mistral-large | mistralai/mistral-large |
Mistral AI | $0.003 | $0.01 | 131,072 | • Model card • Blog post for Mistral Large 2 |
mistral-small-24b-instruct-2501 | mistralai/mistral-small-24b-instruct-2501 |
Mistral AI | $0.00035 | $0.00035 | 32,768 | • Model card • Blog post for Mistral Small 3 |
mixtral-8x7b-instruct-v01 | mistralai/mixtral-8x7b-instruct-v01 |
Mistral AI | $0.0006 | $0.0006 | 32,768 | • Model card • Research paper |
mt0-xxl-13b | bigscience/mt0-xxl |
BigScience | $0.0018 | $0.0018 | 4,096 | • Model card • Research paper |
pixtral-12b | mistralai/pixtral-12b |
Mistral AI | $0.00035 | $0.00035 | 128,000 | • Model card • Blog post for Pixtral 12B |
Deploy on demand foundation models
You can work with a foundation model from a set of IBM-curated models to deploy for the exclusive use of your organization.
IBM deploy on demand foundation models
The following table lists the IBM foundation models that are available for you to deploy on demand.
Some IBM foundation models are also available from third-party repositories, such as Hugging Face. IBM foundation models that you obtain from a third-party repository are not indemnified by IBM. Only IBM foundation models that you access from watsonx.ai are indemnified by IBM. For more information about contractual protections related to IBM indemnification, see the IBM Client Relationship Agreement and IBM watsonx.ai service description.
Model name | Price per hour in USD | Model hosting category | Context window (input + output tokens) |
---|---|---|---|
granite-7b-lab | $5.22 | Small | 4,096 |
granite-13b-chat-v2 | $5.22 | Small | 8,192 |
granite-13b-instruct-v2 | $5.22 | Small | 8,192 |
granite-3b-code-instruct | $5.22 | Small | 128,000 |
granite-8b-code-instruct | $5.22 | Small | 128,000 |
granite-20b-code-instruct | $5.22 | Small | 8,192 |
granite-34b-code-instruct | $5.22 | Small | 8,192 |
granite-20b-code-base-schema-linking | $5.22 | Small | 8,192 |
granite-20b-code-base-sql-gen | $5.22 | Small | 8,192 |
granite-3-8b-base | $5.22 | Small | 4,096 |
Third-party deploy on demand foundation models
The following table lists the third-party foundation models that are available for you to deploy on demand.
Model name | Provider | Price per hour in USD | Model hosting category | Context window (input + output tokens) |
---|---|---|---|---|
deepseek-r1-distill-llama-8b | DeepSeek AI | $5.22 | Small | 131,072 |
deepseek-r1-distill-llama-70b | DeepSeek AI | $20.85 | Large | 131,072 |
flan-t5-xl-3b | $5.22 | Small | 4,096 | |
flan-t5-xxl-11b | $5.22 | Small | 4,096 | |
flan-ul2-20b | $5.22 | Small | 4,096 | |
llama-2-13b-chat | Meta | $5.22 | Small | 4,096 |
llama-2-70b-chat | Meta | $20.85 | Large | 4,096 |
llama-3-8b-instruct | Meta | $5.22 | Small | 8,192 |
llama-3-70b-instruct | Meta | $20.85 | Large | 8,192 |
llama-3-1-8b | Meta | $5.22 | Small | 131,072 |
llama-3-1-8b-instruct | Meta | $5.22 | Small | 131,072 |
llama-3-1-70b-instruct | Meta | $20.85 | Large | 131,072 |
llama-3-3-70b-instruct | Meta | $10.40 | Medium | 131,072 |
llama-3-3-70b-instruct-hf | Meta | $20.85 | Large | 131,072 |
mixtral-8x7b-base | Mistral AI | $10.40 | Medium | 32,768 |
mixtral-8x7b-instruct-v01 | Mistral AI | $10.40 | Medium | 32,768 |
mistral-large-instruct-2407 | Mistral AI | $55.15 (See note.) | Large | 131,072 |
mistral-large-instruct-2411 | Mistral AI | $55.15 (See note.) | Large | 131,072 |
mistral-nemo-instruct-2407 | Mistral AI | $5.22 | Small | 131,072 |
mt0-xxl-13b | BigScience | $5.22 | Small | 4,096 |
Hosting: $20.85 + Access: $34.30 = Total: $55.15 USD per hour
Learn more
- IBM foundation models
- Third-party foundation models
- For more information about the foundation models that IBM provides for embedding and reranking text, see Supported encoder models.
- For a list of which models are provided in each regional data center, see Regional availability of foundation model.
- For details about foundation model pricing, Billing details for generative AI assets
- For information about pricing and rate limiting, see watsonx.ai Runtime plans.
Parent topic: Gen AI solutions