Amazon Bedrock Model Catalog 2026

First Published:
Last Updated:

* The contents of this article reflect the state as of the Last Updated date shown above. Bedrock catalog rows are revised in place as new models GA or are retired; check the model card before any production decision.

Snapshot scope: This is a point-in-time reference of every foundation model offered as a Bedrock-managed inference target.
Last reviewed: 2026-05.
Update cadence: refreshed as needed within the publication year; the URL stays stable across refreshes and each refresh overwrites this page in place.
Sources of truth: AWS Bedrock User Guide "Models at a glance", per-model cards, and the Regional Availability pages.

In late 2023 Amazon Bedrock launched with a single-digit number of model families. As of 2026, the catalog spans 18 providers and well over 110 individually addressable model variants, including 40+ language and reasoning models, three monolithic image generators plus 13 image-editing primitives from Stability AI, multiple speech and video models, and three inference routing tiers (In-Region, Geo, Global). Almost every model now ships with at least one cross-Region inference profile, so picking a model is no longer just "name plus Region" but "name plus Region plus inference profile prefix".

For engineers, architects, and PoC leads, the practical question is no longer "is the model on Bedrock?" but "which exact ID do I call, in which Region, with which capabilities switched on?" This page is built to answer that in one fetch.

1. How This Snapshot Was Compiled

1.1 Scope and Inclusion Criteria

This page covers every foundation model that is callable through the Amazon Bedrock bedrock-runtime inference endpoint as of the review date in the header. We include:
  • Generally Available (GA) and Active lifecycle models.
  • Models in Legacy lifecycle that still accept inference calls but have a published End-of-Life (EOL) date.
  • Preview models that are visible in the public model-cards index and accept inference (these are tagged [Preview]).

We exclude:
  • Models that exist only in the Bedrock Marketplace as third-party listings without a Bedrock-managed inference endpoint.
  • Models in Imported model form that customers bring themselves via Custom Model Import.
  • Anything restricted to non-public allowlists.

1.2 Sources of Truth

Every row in this catalog is cross-checked against three AWS sources:

Pricing is deliberately not included on this page. Bedrock token-rate pricing changes more frequently than this snapshot is regenerated, and stale numbers create more risk than value. For pricing, follow the Amazon Bedrock Pricing page.

1.3 Update Cadence

This page is refreshed as needed within the publication year. We retain the same URL on each refresh so that links from other pages, AI agents, and search engines do not rot. The Last reviewed date in the header is the single source of truth for "how fresh is this?".

1.4 Notation: Badges and Symbols

Throughout the tables we use the following compact notation:
  • [Preview] — the model is publicly visible but not yet GA. Quota and SLA may differ.
  • [Legacy] — the model is still callable but the provider has scheduled a successor; an EOL date is published.
  • [EOL YYYY-MM-DD] — the model will stop accepting inference on this date.
  • Yes / - — feature supported / not supported.
  • In-Region / Geo / Global — which inference routing tier(s) the model supports in a given Region (see §4.4).
  • n/a — the column does not apply (for example, Output length on an embedding model, or Context on a small on-demand LLM that is sized by request rather than by a published context cap).

A note on Model IDs: Bedrock uses several ID conventions in parallel and the differences are not cosmetic - the exact string is required by InvokeModel / Converse. Patterns to watch for:
  • Dated and revision-suffixed IDs: anthropic.claude-haiku-4-5-20251001-v1:0, mistral.pixtral-large-2502-v1:0, twelvelabs.pegasus-1-2-v1:0.
  • Clean suffix-less IDs for newer releases: anthropic.claude-opus-4-7, anthropic.claude-sonnet-4-6, mistral.devstral-2-123b, qwen.qwen3-vl-235b-a22b.
  • Dot-separated version numbers: zai.glm-4.7 (not zai.glm-4-7), minimax.minimax-m2.5, moonshotai.kimi-k2.5, deepseek.v3.2.
  • Double-namespace IDs: minimax.minimax-m2; same provider name appears in both the provider segment and the model name.
  • Two provider prefixes for one provider: Moonshot publishes Kimi K2 Thinking as moonshot.kimi-k2-thinking and Kimi K2.5 as moonshotai.kimi-k2.5. Always copy the exact prefix from the model card.
  • Implicit context-length variants: many models also expose ID aliases that pin the maximum context, e.g. amazon.nova-2-lite-v1:0:256k, amazon.nova-premier-v1:0:1000k, meta.llama3-3-70b-instruct-v1:0:128k. These behave like the base ID but with a hard context cap; the catalog tables list the unsuffixed base ID.

2. Quick Reference Summary

2.1 Model Count by Provider

The headline number for 2026 is that Bedrock now exposes models from 18 providers, a sharp jump from the 6 providers that defined the 2024 snapshot. Many of the additions are open-weight providers (Google Gemma, OpenAI's GPT-OSS family, Qwen, DeepSeek, Z.AI, Moonshot, MiniMax, NVIDIA Nemotron) whose models are hosted on Bedrock as Bedrock-managed inference rather than as fully proprietary APIs. Counts below include Active and Legacy GA models that still accept inference calls, but exclude older revision aliases such as ...-v1:0:200k.

* You can sort the table by clicking on the column name.
ProviderGA models in catalog
Stability AI16
Amazon16
Mistral AI14
Meta12
Anthropic11
Qwen7
Cohere6
OpenAI4
NVIDIA4
Z.AI3
MiniMax3
Google3
Writer3
TwelveLabs3
DeepSeek3
Moonshot AI2
AI21 Labs2
Luma AI1
Total113

2.2 Model Count by Modality

* You can sort the table by clicking on the column name.
ModalityCount (approx.)Examples
Text generation (chat / reasoning / coding)70+Claude Opus 4.7, Nova 2 Lite, Llama 4 Maverick, Mistral Large 3, Qwen3 235B A22B 2507, GLM 5
Vision (image input → text output)20+Claude 4.x family, Nova Pro / Premier / 2 Lite, Llama 3.2 Vision (Legacy) / Llama 4, Pixtral Large, Magistral Small 2509, Qwen3 VL, Gemma 3, Writer Palmyra Vision 7B
Image generation and editing18Stability AI suite (16: SD3.5 Large, Stable Image Core / Ultra, plus 13 editing primitives), Amazon Nova Canvas (Legacy), Titan Image Generator G1 v2 (Legacy)
Video understanding4Nova Premier, Nova 2 Lite, TwelveLabs Pegasus v1.2, TwelveLabs Marengo Embed (video input)
Video generation3Amazon Nova Reel (Legacy), Luma Ray v2, (Nova Reel earlier v1:0 build also still Legacy-listed)
Speech-to-speech (bidirectional)2Amazon Nova Sonic (Legacy), Amazon Nova 2 Sonic
Audio in → Text out2Mistral Voxtral Mini 3B 2507, Voxtral Small 24B 2507
Text embeddings6Cohere Embed (English v3 / Multilingual v3 / v4), Amazon Titan Text Embeddings V2, Amazon Titan Embeddings G1 - Text v1, Amazon Titan Embeddings G1 - Text v2 (amazon.titan-embed-g1-text-02)
Multimodal embeddings4Amazon Nova 2 Multimodal Embeddings, Amazon Titan Multimodal Embeddings G1, TwelveLabs Marengo Embed v2.7, TwelveLabs Marengo Embed 3.0
Rerank2Cohere Rerank 3.5, Amazon Rerank

2.3 Headline Numbers

  • Largest context window: 1,000,000 tokens (1M), shared by Anthropic Claude Opus 4.7, Amazon Nova Premier (Legacy), Amazon Nova 2 Lite, and Meta Llama 4 Maverick / Scout.
  • Largest output: 128K tokens, on Anthropic Claude Opus 4.7. The rest of the Claude 4.x family clusters at 64K output (32K on the original Claude Opus 4, now Legacy).
  • AWS Regions with at least one Bedrock model: 30+ commercial Regions when Global cross-Region inference is considered; far fewer (typically 4–8) when you constrain to In-Region inference for data residency.
  • Inference profile prefixes in use: us., eu., apac., jp., au., and global. (see §4.4).
  • Latest GA at review time: Claude Opus 4.7 (2026-04-16) and the Nova 2 generation (Nova 2 Lite, Nova 2 Sonic, Nova 2 Multimodal Embeddings).

3. Detailed Model Catalog

3.1 Bedrock Model Family Lineage

The catalog is wide enough that a single index helps orient the rest of this page. The table below groups models by provider and family axis (text vs. multimodal vs. image vs. speech vs. embedding/rerank).

* You can sort the table by clicking on the column name.
ProviderFamilyExample modelsModality
AnthropicClaude 4.xOpus 4.7 / 4.6 / 4.5 / 4.1, Opus 4 [Legacy], Sonnet 4.6 / 4.5, Sonnet 4 [Legacy], Haiku 4.5Text + Vision
AnthropicClaude 3.x3.5 Haiku [Legacy], 3 Haiku [Legacy]Text + Vision
AmazonNova 2Lite, Sonic, Multimodal EmbeddingsMultimodal / Speech / Embedding
AmazonNova (Gen 1)Premier [Legacy], Pro, Lite, MicroMultimodal
AmazonNova StudioCanvas [Legacy], Reel [Legacy], Sonic [Legacy]Image / Video / Speech
AmazonTitanImage Gen G1 v2 [Legacy], Text Embeddings V2, MM Embeddings G1, Embed G1 Text v1 / v2, RerankImage / Embedding / Rerank
MetaLlama 4Maverick 17B, Scout 17BMultimodal MoE
MetaLlama 3.x3.3 70B Instruct; 3.2 1B / 3B / 11B Vision / 90B Vision [all Legacy]; 3.1 8B / 70B, 3.1 405B [Legacy]; 3 8B / 70B InstructText / Vision
Mistral AIMistral Large / Small / Mixtral / PixtralLarge 3 (675B), Large 24.07, Large 24.02, Small, 7B, Mixtral 8x7B, Pixtral Large 25.02Text / MoE / Multimodal
Mistral AIMinistral 3 / Voxtral / Devstral / MagistralMinistral 3 (3B / 8B / 14B), Voxtral Mini 3B / Small 24B, Devstral 2 123B, Magistral Small 2509Text small / Audio / Code / Reasoning + Vision
CohereCommand / Embed / RerankCommand R+ [Legacy], Command R [Legacy], Embed English v3 / Multilingual v3 / v4, Rerank 3.5Text / Embedding / Rerank
AI21 LabsJamba 1.5Large, MiniHybrid SSM
Stability AIStable Image (monolithic + editing)SD3.5 Large, Stable Image Core / Ultra, plus 13 editing primitives (Conservative / Creative / Fast Upscale, Inpaint, Outpaint, Erase Object, Remove Background, Control Sketch / Structure, Search Recolor / Replace, Style Guide / Transfer)Image generation / editing
DeepSeekV3 / R1V3.2, V3, R1Text / Reasoning
GoogleGemma 327B IT, 12B IT, 4B ITText + Vision (open weight)
MiniMaxM2M2.5, M2.1, M2Text + Multimodal
Moonshot AIKimiK2.5, K2 ThinkingText / Reasoning
NVIDIANemotronSuper 3 120B A12B, Nano 3 30B, Nano 12B v2, Nano 9B v2Text
OpenAIGPT-OSSgpt-oss-120b, gpt-oss-20b, Safeguard 120B, Safeguard 20BText (open weight) / Classification
QwenQwen3Qwen3 235B A22B 2507, Qwen3 VL 235B A22B, Qwen3 Next 80B A3B, Qwen3 32B, Coder Next / 480B A35B / 30B A3BText / Vision / Code (MoE)
WriterPalmyraPalmyra X5, X4, Vision 7BText / Vision
TwelveLabsMarengo / PegasusMarengo Embed 3.0, Marengo Embed v2.7, Pegasus v1.2Video understanding / embedding
Z.AIGLMGLM 5, GLM 4.7, GLM 4.7 FlashText
Luma AIRayRay v2Video generation

3.2 Provider-Indexed Anchor List

Jump directly to a provider:

3.3 Anthropic - Claude Family

The Claude family on Bedrock spans the GA Claude 4.x generation and a Claude 3 tail that has now largely transitioned to the Legacy lifecycle. Claude Opus 4.7 is the headline model at the review date, launched on 2026-04-16 with a 1M-token context window and 128K output. All Claude 4.x models support Tool Use, Prompt Caching, Guardrails, and Knowledge Bases.

Note on ID conventions: in 2026 Bedrock uses two ID forms for Claude in parallel. The newest releases (Opus 4.7, Opus 4.6, Sonnet 4.6) ship with cleaner suffix-less IDs; older releases keep dated, versioned IDs such as anthropic.claude-opus-4-5-20251101-v1:0. Both forms are accepted by InvokeModel / Converse.

* You can sort the table by clicking on the column name.
Model nameModel ID (bedrock-runtime)ModalityContextOutputStreamingTool UsePrompt CachingCross-Region InferenceLifecycle
Claude Opus 4.7anthropic.claude-opus-4-7Text + Vision1M128KYesYesYes (1K min, 4 ckpt, 5 min TTL)us, eu, jp, au, globalActive
Claude Opus 4.6anthropic.claude-opus-4-6-v1Text + Vision200K64KYesYesYesus, eu, au, globalActive
Claude Opus 4.5anthropic.claude-opus-4-5-20251101-v1:0Text + Vision200K64KYesYesYesus, eu, globalActive
Claude Opus 4.1anthropic.claude-opus-4-1-20250805-v1:0Text + Vision200K64KYesYesYesusActive
Claude Opus 4anthropic.claude-opus-4-20250514-v1:0Text + Vision200K32KYesYesYesusLegacy
Claude Sonnet 4.6anthropic.claude-sonnet-4-6Text + Vision200K64KYesYesYesus, eu, jp, au, globalActive
Claude Sonnet 4.5anthropic.claude-sonnet-4-5-20250929-v1:0Text + Vision200K64KYesYesYesus, eu, jp, au, globalActive
Claude Sonnet 4anthropic.claude-sonnet-4-20250514-v1:0Text + Vision200K64KYesYesYesus, eu, apac, globalLegacy
Claude Haiku 4.5anthropic.claude-haiku-4-5-20251001-v1:0Text + Vision200K64KYesYesYes (4K min, 4 ckpt, 5 min and 1 hour TTL)us, eu, jp, globalActive
Claude 3.5 Haikuanthropic.claude-3-5-haiku-20241022-v1:0Text + Vision200K8KYesYesYesusLegacy
Claude 3 Haikuanthropic.claude-3-haiku-20240307-v1:0Text + Vision200K4KYesYes-us, eu, apacLegacy

Note: Starting with Claude Opus 4.7, the temperature, top_p, and top_k sampling parameters are no longer supported — guide model behavior via prompting instead. Extended thinking on Opus 4.7 is thinking.type: "adaptive" only; the older thinking.type: "enabled" with budget_tokens returns a 400 error.

3.4 Amazon - Nova / Titan / Rerank

Amazon's own portfolio has the broadest modality span on Bedrock. Nova 2 (Lite, Sonic, Multimodal Embeddings) is the current generation, launched in late 2025 and early 2026. The first-generation Nova lineup (Premier, Pro, Lite, Micro, the original Sonic / Canvas / Reel) and the older Titan Image Generator have all transitioned to Legacy lifecycle as their Nova 2 successors GA. The Titan embedding models still anchor the long-tail of RAG workloads.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextOutputTool UsePrompt CachingCross-Region InferenceLifecycle
Nova 2 Liteamazon.nova-2-lite-v1:0Text + Vision + Video1M (256K variant also published as amazon.nova-2-lite-v1:0:256k)64KYesYes (1K min, 4 ckpt, 5 min)us, eu, jp, globalActive
Nova 2 Sonicamazon.nova-2-sonic-v1:0Speech in → Speech + Text out (bidirectional)n/an/an/a-In-Region (us-east-1, us-west-2, ap-northeast-1)Active
Nova 2 Multimodal Embeddingsamazon.nova-2-multimodal-embeddings-v1:0Embedding (text + image + video)n/an/an/a-In-Region (us-east-1)Active
Nova Premieramazon.nova-premier-v1:0Text + Vision + Video1M25KYesYes (1K min, 4 ckpt, 5 min)usLegacy (EOL 2026-09-14)
Nova Proamazon.nova-pro-v1:0Text + Vision + Video300K5KYesYesus, eu, apacActive
Nova Liteamazon.nova-lite-v1:0Text + Vision + Video300K5KYesYesus, eu, apacActive
Nova Microamazon.nova-micro-v1:0Text128K5KYesYesus, eu, apacActive
Nova Sonicamazon.nova-sonic-v1:0Speech in → Speech + Text outn/an/an/a-In-Region (us-east-1, ap-northeast-1)Legacy (succeeded by Nova 2 Sonic)
Nova Canvasamazon.nova-canvas-v1:0Image generation (text + image input)n/an/an/a-In-Region (us-east-1, ap-northeast-1, eu-west-1)Legacy (EOL 2026-09-30 in us-east-1)
Nova Reelamazon.nova-reel-v1:1Video generation (text + image input)n/an/an/a-In-Region (us-east-1, ap-northeast-1, eu-west-1)Legacy (a prior nova-reel-v1:0 build is also still listed as Legacy)
Titan Image Generator G1 v2amazon.titan-image-generator-v2:0Image generation (text + image input)n/an/an/a-In-Region (us-east-1, us-west-2)Legacy
Titan Multimodal Embeddings G1amazon.titan-embed-image-v1Embedding (text + image)n/an/an/a-In-Region (us-east-1, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-central-1, eu-west-1)Active
Titan Text Embeddings V2amazon.titan-embed-text-v2:0Embedding (text)8K inputn/an/a-In-Region in nearly every commercial RegionActive
Titan Embeddings G1 - Text v1amazon.titan-embed-text-v1Embedding (text)8K inputn/an/a-In-Region (us-east-1, us-west-2, ap-northeast-1, eu-central-1, eu-west-1)Active
Titan Embeddings G1 - Text v2 (alt build)amazon.titan-embed-g1-text-02Embedding (text)8K inputn/an/a-In-Region (us-east-1, us-west-2, eu-central-1)Active
Rerankamazon.rerank-v1:0Rerankn/an/an/a-In-Region (us-west-2, ap-northeast-1, eu-central-1)Active

3.5 Meta - Llama Family

Llama 4 introduced mixture-of-experts to the Bedrock-hosted Llama line. Maverick and Scout are both 17B-active MoEs with a 1M context window and multimodal (image) input. Llama 3.x covers a wider parameter range, including the dense 405B model in 3.1, though the entire 3.2 sub-family and Llama 3.1 405B have transitioned to Legacy as their successors stabilized.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextOutputTool UseCross-Region InferenceLifecycle
Llama 4 Maverick 17B Instructmeta.llama4-maverick-17b-instruct-v1:0Text + Vision (MoE)1M8KYesusActive
Llama 4 Scout 17B Instructmeta.llama4-scout-17b-instruct-v1:0Text + Vision (MoE)1M8KYesusActive
Llama 3.3 70B Instructmeta.llama3-3-70b-instruct-v1:0Text128K8KYesusActive
Llama 3.2 90B Instruct (Vision)meta.llama3-2-90b-instruct-v1:0Text + Vision128K8KYesusLegacy
Llama 3.2 11B Instruct (Vision)meta.llama3-2-11b-instruct-v1:0Text + Vision128K8KYesusLegacy
Llama 3.2 3B Instructmeta.llama3-2-3b-instruct-v1:0Text128K8KYesus, euLegacy
Llama 3.2 1B Instructmeta.llama3-2-1b-instruct-v1:0Text128K8KYesus, euLegacy
Llama 3.1 405B Instructmeta.llama3-1-405b-instruct-v1:0Text128K8KYesusLegacy
Llama 3.1 70B Instructmeta.llama3-1-70b-instruct-v1:0Text128K8KYesusActive
Llama 3.1 8B Instructmeta.llama3-1-8b-instruct-v1:0Text128K8KYesusActive
Llama 3 70B Instructmeta.llama3-70b-instruct-v1:0Text8K8K-In-Region (us-east-1, us-west-2, ap-south-1)Active
Llama 3 8B Instructmeta.llama3-8b-instruct-v1:0Text8K8K-In-Region (us-east-1, us-west-2, ap-south-1)Active

3.6 Mistral AI

Mistral's Bedrock catalog now spans the dense Mistral Large 3 flagship (published with the explicit -675b-instruct suffix in the Bedrock model ID), a classic mixture-of-experts model (Mixtral 8x7B), the multimodal Pixtral Large, the Ministral 3 small-LLM line (3B / 8B / 14B), audio-capable Voxtral models, the Devstral coding model, and Magistral Small for reasoning. Magistral Small 2509 accepts image input in addition to text, and the Voxtral models accept speech alongside text input.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextTool UseCross-Region InferenceLifecycle
Mistral Large 3mistral.mistral-large-3-675b-instructText128KYesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Mistral Large 24.07mistral.mistral-large-2407-v1:0Text128KYesIn-Region (us-west-2)Active
Mistral Large 24.02mistral.mistral-large-2402-v1:0Text32KYesIn-Region (us-east-1, us-west-2, ap-southeast-2, ap-south-1, eu-west-1)Active
Mistral Smallmistral.mistral-small-2402-v1:0Text32KYesIn-Region (us-east-1)Active
Mistral 7B Instructmistral.mistral-7b-instruct-v0:2Text32K-In-Region (us-east-1, us-west-2, ap-southeast-2, ap-south-1, eu-west-1)Active
Mixtral 8x7B Instructmistral.mixtral-8x7b-instruct-v0:1Text (MoE)32K-In-Region (us-east-1, us-west-2, ap-southeast-2, ap-south-1, eu-west-1)Active
Pixtral Large 25.02mistral.pixtral-large-2502-v1:0Text + Vision128KYesus, euActive
Ministral 3 14B Instructmistral.ministral-3-14b-instructTextn/a (small LLM, On-Demand)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Ministral 3 8B Instructmistral.ministral-3-8b-instructTextn/a (small LLM, On-Demand)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Ministral 3 3B Instructmistral.ministral-3-3b-instructTextn/a (small LLM, On-Demand)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Devstral 2 123Bmistral.devstral-2-123bText (coding)n/a (On-Demand)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-central-1, eu-west-1)Active
Magistral Small 2509mistral.magistral-small-2509Text + Vision (reasoning)n/a (On-Demand)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Voxtral Mini 3B 2507mistral.voxtral-mini-3b-2507Speech + Text in → Text outn/a (On-Demand)-In-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active
Voxtral Small 24B 2507mistral.voxtral-small-24b-2507Speech + Text in → Text outn/a (On-Demand)-In-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-west-1)Active

3.7 Cohere

The Cohere Command R / R+ generative pair has now transitioned to Legacy as customers migrate to Embed v4 plus stronger general-purpose models from other providers; the Embed and Rerank lines remain Active and form the backbone of many RAG stacks on Bedrock.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextTool UseCross-Region InferenceLifecycle
Command R+cohere.command-r-plus-v1:0Text128KYesIn-Region (us-east-1, us-west-2)Legacy
Command Rcohere.command-r-v1:0Text128KYesIn-Region (us-east-1, us-west-2)Legacy
Embed English v3cohere.embed-english-v3Embeddingn/an/aIn-Region (us-east-1, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-central-1, eu-west-1)Active
Embed Multilingual v3cohere.embed-multilingual-v3Embeddingn/an/aIn-Region (us-east-1, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1, eu-central-1, eu-west-1)Active
Embed v4cohere.embed-v4:0Embeddingn/an/aus, eu, global (Geo + Global available)Active
Rerank 3.5cohere.rerank-v3-5:0Rerankn/an/aIn-Region (us-west-2, ap-northeast-1, eu-central-1)Active

3.8 AI21 Labs - Jamba

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextTool UseCross-Region InferenceLifecycle
Jamba 1.5 Largeai21.jamba-1-5-large-v1:0Text (Hybrid SSM/Transformer)256KYesIn-Region (us-east-1)Active
Jamba 1.5 Miniai21.jamba-1-5-mini-v1:0Text (Hybrid SSM/Transformer)256KYesIn-Region (us-east-1)Active

3.9 Stability AI - Image Generation

Stability's 2026 catalog is built around two layers: a small set of monolithic text-to-image generators (Stable Diffusion 3.5 Large, Stable Image Core, Stable Image Ultra) plus a wide set of editing primitives where each model handles one operation (inpaint, outpaint, upscale, control, search-and-replace, style). Each model accepts text and/or image inputs and returns an image. The editing primitives are hosted in us-west-2 with a subset also in us-east-1; the three monolithic generators are us-west-2-only at the review date. US Geo cross-Region profiles (us.stability....) are available for the editing primitives.

Naming convention note: the Bedrock model IDs are mostly prefixed stability.stable-image-..., but five IDs are published without the image- infix: stability.stable-conservative-upscale-v1:0, stability.stable-creative-upscale-v1:0, stability.stable-fast-upscale-v1:0, stability.stable-outpaint-v1:0, and stability.stable-style-transfer-v1:0. Always copy the exact ID from the model card before wiring it into IaC.

* You can sort the table by clicking on the column name.
Model nameModel IDOperation
Stable Diffusion 3.5 Largestability.sd3-5-large-v1:0Text-to-image (monolithic, high quality)
Stable Image Corestability.stable-image-core-v1:1Text-to-image (monolithic, fast)
Stable Image Ultrastability.stable-image-ultra-v1:1Text-to-image (monolithic, ultra quality)
Stable Image Conservative Upscalestability.stable-conservative-upscale-v1:0Upscale (faithful)
Stable Image Creative Upscalestability.stable-creative-upscale-v1:0Upscale (creative)
Stable Image Fast Upscalestability.stable-fast-upscale-v1:0Upscale (fast)
Stable Image Inpaintstability.stable-image-inpaint-v1:0Inpaint
Stable Image Outpaintstability.stable-outpaint-v1:0Outpaint
Stable Image Erase Objectstability.stable-image-erase-object-v1:0Object erase
Stable Image Remove Backgroundstability.stable-image-remove-background-v1:0Background remove
Stable Image Control Sketchstability.stable-image-control-sketch-v1:0Sketch-guided gen
Stable Image Control Structurestability.stable-image-control-structure-v1:0Structure-guided gen
Stable Image Search and Recolorstability.stable-image-search-recolor-v1:0Object recolor
Stable Image Search and Replacestability.stable-image-search-replace-v1:0Object replace
Stable Image Style Guidestability.stable-image-style-guide-v1:0Style transfer (guide)
Stable Image Style Transferstability.stable-style-transfer-v1:0Style transfer (full)

3.10 DeepSeek

DeepSeek is on Bedrock as Bedrock-managed inference behind US-prefixed inference profiles. The Bedrock model IDs use dot-versioned naming (deepseek.v3.2) for the latest revision and the conventional -v1:0 suffix for the earlier ones.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityContextTool UseCross-Region InferenceLifecycle
DeepSeek V3.2deepseek.v3.2Text128K (typical for V3 family)YesIn-Region (us-east-1, us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1)Active
DeepSeek V3deepseek.v3-v1:0Text128KYesIn-Region (us-east-2, us-west-2, ap-northeast-1, ap-southeast-2, ap-south-1)Active
DeepSeek-R1deepseek.r1-v1:0Text (reasoning)128KYesus (Geo profile only; us.deepseek.r1-v1:0)Active

3.11 Google - Gemma

Google Gemma 3 ships on Bedrock as open-weight chat-tuned models. All three sizes are instruction-tuned (IT); no pretrained-only (PT) build is exposed through Bedrock-managed inference at the review date. All three accept image input alongside text.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
Gemma 3 27B ITgoogle.gemma-3-27b-itText + VisionActive
Gemma 3 12B ITgoogle.gemma-3-12b-itText + VisionActive
Gemma 3 4B ITgoogle.gemma-3-4b-itText + VisionActive

3.12 MiniMax

The MiniMax M2 family uses double-namespace IDs (minimax.minimax-...) on Bedrock; the version number is dot-separated rather than dash-separated. All three are Active.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
MiniMax M2.5minimax.minimax-m2.5Text + MultimodalActive
MiniMax M2.1minimax.minimax-m2.1Text + MultimodalActive
MiniMax M2minimax.minimax-m2Text + MultimodalActive

3.13 Moonshot AI - Kimi

Moonshot publishes Kimi K2 Thinking under the moonshot. provider namespace, while Kimi K2.5 was added later under the moonshotai. namespace and with dot-separated versioning. Both forms are valid Bedrock IDs - this is a deliberate inconsistency you must respect verbatim.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
Kimi K2.5moonshotai.kimi-k2.5TextActive
Kimi K2 Thinkingmoonshot.kimi-k2-thinkingText (reasoning)Active

3.14 NVIDIA - Nemotron

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
NVIDIA Nemotron 3 Super 120B A12Bnvidia.nemotron-super-3-120bTextActive
Nemotron Nano 3 30Bnvidia.nemotron-nano-3-30bTextActive
Nemotron Nano 12B v2nvidia.nemotron-nano-12b-v2TextActive
NVIDIA Nemotron Nano 9B v2nvidia.nemotron-nano-9b-v2TextActive

3.15 OpenAI - GPT-OSS

OpenAI's open-weight GPT-OSS models arrived on Bedrock in 2025, hosted as Bedrock-managed inference. The Safeguard variants are tuned for content classification rather than generation. The base GPT-OSS IDs carry an explicit -1:0 revision suffix; the Safeguard IDs do not.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
gpt-oss-120bopenai.gpt-oss-120b-1:0Text (open weight)Active
gpt-oss-20bopenai.gpt-oss-20b-1:0Text (open weight)Active
GPT OSS Safeguard 120Bopenai.gpt-oss-safeguard-120bText (classification)Active
GPT OSS Safeguard 20Bopenai.gpt-oss-safeguard-20bText (classification)Active

3.16 Qwen

Qwen3 on Bedrock uses two ID conventions in parallel: the larger newest releases (qwen3-vl-235b-a22b, qwen3-next-80b-a3b, qwen3-coder-next) ship without a -v1:0 suffix, while the dated and Instruct variants carry the conventional -v1:0.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
Qwen3 235B A22B 2507qwen.qwen3-235b-a22b-2507-v1:0Text (MoE)Active
Qwen3 VL 235B A22Bqwen.qwen3-vl-235b-a22bText + Vision (MoE)Active
Qwen3 Next 80B A3Bqwen.qwen3-next-80b-a3bText (MoE)Active
Qwen3 Coder Nextqwen.qwen3-coder-nextText (coding)Active
Qwen3 Coder 480B A35Bqwen.qwen3-coder-480b-a35b-v1:0Text (coding, MoE)Active
Qwen3 Coder 30B A3Bqwen.qwen3-coder-30b-a3b-v1:0Text (coding)Active
Qwen3 32Bqwen.qwen3-32b-v1:0TextActive

3.17 Writer - Palmyra

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
Palmyra X5writer.palmyra-x5-v1:0TextActive
Palmyra X4writer.palmyra-x4-v1:0TextActive
Palmyra Vision 7Bwriter.palmyra-vision-7bText + VisionActive

3.18 TwelveLabs

Marengo Embed accepts all four modalities (text, image, speech, video) and returns an embedding vector. Pegasus v1.2 takes text + video and returns text (video understanding). All three are accessed through inference profiles (us., eu., global.).

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
Marengo Embed 3.0twelvelabs.marengo-embed-3-0-v1:0Text + Image + Speech + Video → EmbeddingActive
Marengo Embed v2.7twelvelabs.marengo-embed-2-7-v1:0Text + Image + Speech + Video → EmbeddingActive
Pegasus v1.2twelvelabs.pegasus-1-2-v1:0Text + Video → Text (video understanding)Active

3.19 Z.AI - GLM

Z.AI publishes GLM with dot-separated versioning (zai.glm-4.7, not zai.glm-4-7). GLM 4.7 Flash is the only Z.AI model available in eu-central-1 at the review date.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityLifecycle
GLM 5zai.glm-5TextActive
GLM 4.7zai.glm-4.7TextActive
GLM 4.7 Flashzai.glm-4.7-flashText (fast)Active

3.20 Luma AI

Luma joined Bedrock in 2026 with a single video-generation model, currently us-west-2-only.

* You can sort the table by clicking on the column name.
Model nameModel IDModalityCross-Region InferenceLifecycle
Luma Ray v2luma.ray-v2:0Video generation (text + image input)In-Region (us-west-2)Active

4. Region Availability Matrix

Bedrock now offers three inference routing tiers. In-Region keeps the request inside a single Region (strict data residency). Geo routes across Regions within a geography (US, EU, APAC, JP, AU). Global routes anywhere worldwide. Most modern models are available in at least one Geo profile and a Global profile; older models are In-Region only.

The matrix below covers 8 representative Regions out of the 30+ Regions where Bedrock has presence. For the complete list, follow the per-model card link.

Legend: In = In-Region available; Geo = available via a Geo cross-Region profile sourced from this Region; Glb = available via the Global profile sourced from this Region; - = not available.

4.1 Text and Vision Models × 8 Regions

* You can sort the table by clicking on the column name.
Modelus-east-1us-east-2us-west-2ap-northeast-1ap-southeast-2ap-south-1eu-central-1eu-west-1
Claude Opus 4.7In + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + GlbIn + GlbIn + Geo + GlbIn + Geo + Glb
Claude Haiku 4.5In + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + GlbIn + GlbIn + Geo + GlbIn + Geo + Glb
Claude Sonnet 4.6In + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + GlbIn + Geo + GlbIn + Geo + Glb
Claude Sonnet 4.5In + Geo + GlbGlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + GlbIn + Geo + GlbIn + Geo + Glb
Claude Sonnet 4 [Legacy]In + Geo + GlbGlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + Glb
Nova 2 LiteIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + Geo + GlbIn + GlbIn + GlbIn + Geo + GlbIn + Geo + Glb
Nova ProIn + GeoIn + Geo-In + GeoIn + GeoIn + Geo-In
Nova LiteIn + GeoIn + Geo-In + GeoIn + GeoIn + Geo-In
Nova MicroIn + GeoIn + Geo-----In
Llama 4 Maverick 17BIn + Geo (us)In + Geo (us)In + Geo (us)-----
Llama 4 Scout 17BIn + Geo (us)In + Geo (us)In + Geo (us)-----
Llama 3.3 70B InstructIn + Geo (us)In + Geo (us)In + Geo (us)-----
Mistral Large 3 (675B)InInInInInIn-In
Mistral Large 24.07--In-----
Pixtral Large 25.02In + Geo (us)-In + Geo (us)---In + Geo (eu)In + Geo (eu)
Command R+ [Legacy]In-In-----
Jamba 1.5 LargeIn-------
DeepSeek V3.2InInInInInIn--
DeepSeek-R1In + Geo (us)In + Geo (us)In + Geo (us)-----

4.2 Image, Video, and Speech Models × 8 Regions

* You can sort the table by clicking on the column name.
Modelus-east-1us-east-2us-west-2ap-northeast-1ap-southeast-2ap-south-1eu-central-1eu-west-1
Nova Canvas [Legacy]In [EOL 2026-09-30]--In---In
Nova Reel [Legacy]In (v1:1 and v1:0)--In (v1:0)---In (v1:0)
Nova Sonic [Legacy]In--In----
Nova 2 SonicIn-InIn----
Titan Image Generator G1 v2 [Legacy]In-In-----
Stable Image suite *In (editing subset)-In (full; incl. SD3.5 Large / Core / Ultra)-----
Luma Ray v2--In-----

* The Stable Image suite is a family of 16 models on Bedrock: three monolithic text-to-image generators (SD3.5 Large, Stable Image Core, Stable Image Ultra) plus 13 per-operation editing primitives (upscale, inpaint, outpaint, erase, remove background, control, search-and-replace, style). The full 16 are hosted in us-west-2; the editing primitives also have an In-Region presence in us-east-1 and corresponding us.stability.... Geo cross-Region profiles. Per-primitive availability varies - confirm against the Bedrock Console Foundation models → Model access view before designing a workload.


4.3 Embedding and Rerank Models × 8 Regions

* You can sort the table by clicking on the column name.
Modelus-east-1us-east-2us-west-2ap-northeast-1ap-southeast-2ap-south-1eu-central-1eu-west-1
Titan Text Embeddings V2InInInInInInIn-
Titan Embeddings G1 - TextIn--In--InIn
Titan Multimodal Embeddings G1----InInInIn
Nova Multimodal EmbeddingsIn-------
Cohere Embed EnglishIn-InIn--In-
Cohere Embed MultilingualIn-InIn--In-
Cohere Rerank 3.5--InIn--In-
Amazon Rerank--InIn--In-

4.4 Cross-Region Inference Profiles

Cross-Region inference profile IDs use a stable prefix that encodes the destination geography. As of 2026 the following prefixes are in active use:

PrefixGeographyTypical destination Regions
us.<model-id>USus-east-1, us-east-2, us-west-2 (and us-west-1 / ca-central-1 as routed sources)
eu.<model-id>Europeeu-central-1, eu-north-1, eu-south-1, eu-south-2, eu-west-1, eu-west-3
apac.<model-id>Asia Pacific (broad)ap-northeast-1, ap-southeast-1, ap-southeast-2, ap-south-1 (and others)
jp.<model-id>Japanap-northeast-1 (Tokyo), ap-northeast-3 (Osaka)
au.<model-id>Australia / NZap-southeast-2 (Sydney), ap-southeast-4 (Melbourne), ap-southeast-6 (New Zealand)
global.<model-id>WorldwideAll commercial Regions; routes are model-dependent

Two important nuances:
  • The destination set of a Global profile can change over time as AWS adds Regions. The destination set of a Geo profile (us., eu., jp., etc.) is fixed.
  • Some inference profiles route differently depending on the source Region. For example, calling us.anthropic.claude-3-haiku-20240307-v1:0 from us-east-2 can route to us-east-1, us-east-2, or us-west-2; calling it from us-west-2 can only route to us-east-1 or us-west-2. Use GetInferenceProfile from the source Region to confirm.

A representative subset of profile IDs (verified against list-inference-profiles at the review date):

ModelGeo profile ID(s)Global profile ID
Claude Opus 4.7us.anthropic.claude-opus-4-7, eu.anthropic.claude-opus-4-7, jp.anthropic.claude-opus-4-7global.anthropic.claude-opus-4-7
Claude Opus 4.6us.anthropic.claude-opus-4-6-v1, eu.anthropic.claude-opus-4-6-v1, au.anthropic.claude-opus-4-6-v1global.anthropic.claude-opus-4-6-v1
Claude Haiku 4.5us.anthropic.claude-haiku-4-5-20251001-v1:0, eu.anthropic.claude-haiku-4-5-20251001-v1:0, jp.anthropic.claude-haiku-4-5-20251001-v1:0global.anthropic.claude-haiku-4-5-20251001-v1:0
Claude Sonnet 4.6us.anthropic.claude-sonnet-4-6, eu.anthropic.claude-sonnet-4-6, jp.anthropic.claude-sonnet-4-6, au.anthropic.claude-sonnet-4-6global.anthropic.claude-sonnet-4-6
Claude Sonnet 4 [Legacy]us.anthropic.claude-sonnet-4-20250514-v1:0, eu.anthropic.claude-sonnet-4-20250514-v1:0, apac.anthropic.claude-sonnet-4-20250514-v1:0global.anthropic.claude-sonnet-4-20250514-v1:0
Nova Premier [Legacy]us.amazon.nova-premier-v1:0not supported
Nova 2 Liteus.amazon.nova-2-lite-v1:0, eu.amazon.nova-2-lite-v1:0, jp.amazon.nova-2-lite-v1:0global.amazon.nova-2-lite-v1:0
Nova Pro / Lite / Microus.amazon.nova-{pro,lite,micro}-v1:0, eu.amazon.nova-{pro,lite,micro}-v1:0, apac.amazon.nova-{pro,lite,micro}-v1:0not supported
Llama 4 Maverick / Scout 17Bus.meta.llama4-{maverick,scout}-17b-instruct-v1:0not supported
DeepSeek-R1us.deepseek.r1-v1:0not supported
Cohere Embed v4us.cohere.embed-v4:0, eu.cohere.embed-v4:0global.cohere.embed-v4:0
TwelveLabs Marengo Embed 3.0us.twelvelabs.marengo-embed-3-0-v1:0, eu.twelvelabs.marengo-embed-3-0-v1:0not supported
TwelveLabs Pegasus v1.2us.twelvelabs.pegasus-1-2-v1:0, eu.twelvelabs.pegasus-1-2-v1:0global.twelvelabs.pegasus-1-2-v1:0
Mistral Pixtral Largeus.mistral.pixtral-large-2502-v1:0, eu.mistral.pixtral-large-2502-v1:0not supported

Notable absences: as of the review date, Claude Sonnet 4.6 does not ship an apac. profile (only jp. within APAC). The apac. profile family for Anthropic is concentrated on the Claude 3.x line and Claude Sonnet 4 (Legacy); the Sonnet 4.6 / 4.5 generation jumped directly to the more granular jp. / au. profiles.

5. Recent Additions (2026-02 to 2026-05)

The following models were added or promoted to GA on Bedrock in the rolling three-month window preceding the review date:
  • Anthropic Claude Opus 4.7 — GA 2026-04-16. First Bedrock Claude with a 1M context window and 128K max output. Adaptive-only extended thinking; sampling parameters dropped.
  • Amazon Nova 2 Lite — GA 2025-12-02 (carried over into early 2026). The first member of the Nova 2 generation, with multimodal in (text + image + video), 1M context, 64K output, In-Region in all 8 representative Regions, plus us., eu., jp., and Global cross-Region profiles.
  • Amazon Nova 2 Sonic — GA early 2026. Bidirectional speech-in / speech+text-out model; replaces Nova Sonic (now Legacy) for new builds.
  • Amazon Nova 2 Multimodal Embeddings (amazon.nova-2-multimodal-embeddings-v1:0) — new in early 2026, supersedes the original Nova Multimodal Embeddings ID. Currently In-Region (us-east-1) only.
  • Luma AI Ray v2 (luma.ray-v2:0) — brings Luma as the 18th Bedrock provider. Video generation, us-west-2-only at the review date.
  • MiniMax M2.5 / Moonshot Kimi K2.5 / Z.AI GLM 5 / Qwen3 expansion / NVIDIA Nemotron 3 (Super 120B A12B + Nano 3 30B / 9B v2 / 12B v2) — first-half 2026 expansion of open-weight providers on Bedrock-managed inference.
  • Stability AI — the legacy single Stable Diffusion XL model has been replaced by a 16-model lineup: three monolithic generators (Stable Diffusion 3.5 Large, Stable Image Core, Stable Image Ultra) plus 13 per-operation editing primitives.
  • Mistral Large 3 (mistral.mistral-large-3-675b-instruct) — new 675B Mistral flagship; In-Region in 7 of the 8 representative Regions at the review date.

6. Recent Deprecations and Scheduled EOLs

The following models are on a published EOL schedule or have already transitioned to Legacy. If you are calling any of these, plan a migration before the date shown (or before the next refresh in cases where no firm EOL is published yet):
  • Amazon Nova Premier — Lifecycle: Legacy. EOL: 2026-09-14. Successor: Nova 2 Lite (and a future Nova 2 Pro / Premier when GA). Same 1M context, broader Region availability, and Global routing.
  • Amazon Nova Canvas in us-east-1 — Legacy, EOL: 2026-09-30. The Tokyo (ap-northeast-1) and Ireland (eu-west-1) deployments are also Legacy. Plan to migrate to a successor image model (Stability AI Stable Image Ultra / Core / SD3.5 Large are the natural Bedrock-native successors).
  • Amazon Nova Reel (both v1:0 and v1:1 builds) — Legacy across all Regions; nova-reel-v1:0 in us-east-1 has the firm EOL 2026-09-30. Successor for video generation: Luma Ray v2 (luma.ray-v2:0) in us-west-2.
  • Amazon Nova Sonic (Gen 1) — Legacy; successor is Nova 2 Sonic.
  • Amazon Titan Image Generator G1 v2 — Legacy; same successor guidance as Nova Canvas.
  • Anthropic Claude 3.5 Haiku and Claude 3 Haiku — both transitioned to Legacy. The natural successor is Claude Haiku 4.5.
  • Anthropic Claude Opus 4 (20250514) and Claude Sonnet 4 (20250514) — both transitioned to Legacy as the Opus 4.5 / 4.6 / 4.7 and Sonnet 4.5 / 4.6 lineage stabilized.
  • Cohere Command R / Command R+ — both transitioned to Legacy; Cohere's Active surface on Bedrock is now Embed (English v3 / Multilingual v3 / v4) and Rerank 3.5.
  • Meta Llama 3.2 family (1B / 3B / 11B / 90B Vision) and Llama 3.1 405B — all Legacy. Active Meta successors are Llama 4 Maverick / Scout and Llama 3.3 70B Instruct.

Older Bedrock models that have already been removed from the catalog (and are therefore not listed in §3) include the original Claude 2 / 2.1, Claude Instant, Claude 3 Opus, Llama 2 family, Mistral Large v1, the original Stable Diffusion XL v1, and Titan Text G1 - Express / Lite. Note that some legacy Claude 3.x Sonnet IDs (claude-3-sonnet-20240229, claude-3-5-sonnet-20240620, claude-3-5-sonnet-20241022-v2:0, claude-3-7-sonnet-20250219-v1:0) still appear in regional list-foundation-models output for ap-south-1 in Legacy status; only the most current Claude line (Haiku 4.5 + Sonnet 4.x + Opus 4.x) is included in §3.3.

7. Frequently Asked Questions

7.1 What is the largest context window available on Bedrock as of 2026?

1,000,000 tokens (1M). Five models share this: Anthropic Claude Opus 4.7, Amazon Nova Premier (Legacy), Amazon Nova 2 Lite, Meta Llama 4 Maverick, and Meta Llama 4 Scout. For maximum output, Claude Opus 4.7 leads at 128K tokens; the rest of the Active Claude 4.x family clusters at 64K output, with the Legacy Claude Opus 4 (20250514) capped at 32K.

7.2 Which models support Tool Use / Function Calling as of 2026?

Tool Use is supported across essentially the entire 2025-2026 wave of GA models: the full Claude 3.x and 4.x line, all Nova text variants (Premier, Pro, Lite, Micro, Nova 2 Lite), Llama 3.1 and later (including Llama 4 Maverick / Scout), Mistral Large family (Large 3 / 24.07 / 24.02), Pixtral Large, Ministral 3 (3B / 8B / 14B), Devstral 2 123B, Magistral Small 2509, AI21 Jamba 1.5 Large / Mini, DeepSeek-R1 / V3 / V3.2, and most open-weight additions from 2026 (Qwen3, GLM, Kimi K2 Thinking, Nemotron 3 / Nano line). Cohere Command R / R+ also still advertise Tool Use but are now Legacy, so plan a migration. Tool Use is not supported on the older Llama 3, Mistral 7B / Mixtral 8x7B, embedding models, image generation models, speech models, or the Voxtral audio models.

7.3 Which models support Prompt Caching as of 2026?

Prompt caching support is concentrated in the Anthropic and Amazon Nova families:
  • Claude Opus 4.7, Opus 4.6, Opus 4.5, Opus 4.1, Opus 4 (Legacy), Sonnet 4.6, Sonnet 4.5, Sonnet 4 (Legacy), Haiku 4.5 — all support prompt caching. Claude Haiku 4.5 in particular supports both a 5-minute and a 1-hour TTL.
  • Claude 3.5 Haiku (now Legacy) also supports prompt caching.
  • Nova Premier (Legacy), Nova Pro, Nova Lite, Nova Micro, and Nova 2 Lite all support prompt caching. For Nova Premier and Nova 2 Lite the published limits are 1K minimum tokens per checkpoint, up to 4 checkpoints, and a 5-minute TTL (primarily for text prompts); the Pro / Lite / Micro variants are listed as supported without per-checkpoint figures — confirm against the per-model card before sizing a workload.

Other model families generally do not (yet) advertise prompt caching on Bedrock.

7.4 Which models are only available via Cross-Region Inference?

In 2026 the picture is more nuanced than "Geo-only" - most flagship models advertise both In-Region and one or more Geo / Global profiles, and the API call site is what determines whether the request stays inside one Region or routes across a geography. Models that only route via Cross-Region inference profiles at the review date:
  • Anthropic Claude Opus 4.1 (20250805-v1:0) — us. Geo profile only; no In-Region invocation.
  • Amazon Nova Premier [Legacy] — us.amazon.nova-premier-v1:0 only; no Global profile, no In-Region.
  • Meta Llama 4 Maverick / Scout, Llama 3.3 70B, Llama 3.x Instruct familyus. Geo profile only (the model also lists In-Region in the same source Regions for the profile-aware invocation path).
  • DeepSeek-R1us.deepseek.r1-v1:0 only; no Global profile.
  • TwelveLabs Marengo Embed 3.0 / Pegasus v1.2 — Geo profiles (us., eu.) only; Pegasus also has a Global profile.

Models that do support direct In-Region invocation in their flagship Regions include all of the Anthropic Claude Opus 4.5 / 4.6 / 4.7 line, Claude Sonnet 4.5 / 4.6, Claude Haiku 4.5, Nova 2 Lite, Cohere Embed v4, Pixtral Large, Mistral Large 3, Mistral Ministral 3 / Devstral / Magistral / Voxtral, NVIDIA Nemotron, OpenAI GPT-OSS, Qwen3, Z.AI GLM, MiniMax M2 family, Moonshot Kimi, and the Stability AI family.

In contrast, almost all embedding, rerank, image generation, video generation, and speech models are In-Region only and do not offer cross-Region routing. Exceptions to this rule are TwelveLabs Marengo / Pegasus (Geo / Global), Cohere Embed v4 (Geo / Global), and Stability AI's editing primitives (us. Geo only).

7.5 Which models have been deprecated in the most recent refresh?

The notable shifts in the most recent refresh:
  • Anthropic Claude 3.5 Haiku and Claude 3 Haiku moved to Legacy across all Regions where they were previously Active.
  • Anthropic Claude Opus 4 and Claude Sonnet 4 (both 20250514) moved to Legacy as Opus 4.5 / 4.6 / 4.7 and Sonnet 4.5 / 4.6 became the primary release line.
  • Cohere Command R and Command R+ moved to Legacy; Cohere's Active surface is now Embed (v3 English / v3 Multilingual / v4) and Rerank 3.5.
  • Amazon Nova Sonic (Gen 1) moved to Legacy in favor of Nova 2 Sonic.
  • Amazon Nova Canvas / Nova Reel / Titan Image Generator G1 v2 moved to Legacy across all Regions where they exist.
  • Amazon Nova Premier moved to Legacy with a 2026-09-14 EOL.
  • Meta Llama 3.2 (1B / 3B / 11B Vision / 90B Vision) and Llama 3.1 405B moved to Legacy as Llama 4 Maverick / Scout and Llama 3.3 70B took over the Active surface.

If you are calling any of these directly in production, plan migration on the timeline above. The successor for Nova Premier is Nova 2 Lite (and the upcoming Nova 2 Pro / Premier when they GA). The successor for the Nova Canvas / Reel image and video models is Stability AI's image suite plus Luma Ray v2 for video. The successor for the Claude 3 / 3.5 Haiku tail is Claude Haiku 4.5.

8. References

8.1 AWS Documentation (External)


8.2 Internal Articles


8.3 Update Policy

This snapshot is refreshed as needed within the publication year. Last reviewed: 2026-05. The URL is intentionally stable across refreshes — link to it from your own documentation and the link will continue to resolve.

References:
Tech Blog with curated related content

Written by Hidekazu Konishi