Enable Bedrock Model Access
Before deploying Bedrock Chat, enable model access in Amazon Bedrock:- Navigate to the Bedrock Model Access page in us-east-1
- Click Manage model access
- Select all models you wish to use
- Click Save changes
Supported Models
Bedrock Chat supports the following foundation models:Claude Models (Anthropic)
claude-v4-opusclaude-v4.1-opusclaude-v4.5-opusclaude-v4-sonnetclaude-v4.5-sonnetclaude-v4.5-haikuclaude-v3-opusclaude-v3.5-sonnetclaude-v3.5-sonnet-v2claude-v3.7-sonnet(supports extended thinking up to 64k tokens)claude-v3-haikuclaude-v3.5-haiku
Amazon Nova Models
amazon-nova-proamazon-nova-liteamazon-nova-micro
Mistral Models
mistral-7b-instructmixtral-8x7b-instructmistral-largemistral-large-2
DeepSeek Models
deepseek-r1
Meta Llama Models
llama3-3-70b-instructllama3-2-1b-instructllama3-2-3b-instructllama3-2-11b-instructllama3-2-90b-instruct
Other Models
gpt-oss-20bgpt-oss-120b
Global Model Configuration
Control which models appear in the UI for all users.List of model IDs to enable globally. Empty array enables all models. When set, only specified models appear in dropdown menus.
Only models enabled in the Bedrock console AND included in
globalAvailableModels will be available to users.Default Model Selection
Specify which model is pre-selected when users start a new chat.Model ID pre-selected in the chat UI when users start a new conversation.
Title Generation Model
Specify which model automatically generates conversation titles.Model used to generate short conversation titles. Defaults to
defaultModel if not specified, then falls back to claude-v3-haiku.Bedrock Region Configuration
Specify which AWS region to use for Amazon Bedrock API calls.AWS region where Amazon Bedrock APIs are available. Must be a region where Bedrock is supported.
Amazon Bedrock is not available in all AWS regions. Check the Bedrock documentation for supported regions.
Cross-Region and Global Inference
Amazon Bedrock can dynamically route inference requests across regions for improved throughput and resilience.Global Inference
Enable global inference to route requests to the optimal region worldwide based on latency and availability.
Cross-Region Inference
Enable cross-region inference to route requests within the same geographic area (e.g., within US regions).
Some AWS Service Control Policies (SCPs) may restrict global or cross-region inference. Configure these settings based on your organization’s policies.
Model Parameters
Users can configure generation parameters for each conversation:Temperature
- Range: 0 to 1
- Step: 0.05
- Default: 0.6
- Purpose: Controls randomness in responses. Lower values make output more deterministic.
Max Tokens
- Range: 1 to 64,000
- Step: 10
- Default: 2,000
- Purpose: Maximum number of tokens to generate in the response.
Claude 3.7 with extended thinking supports up to 64,000 tokens output.
Top P
- Range: 0 to 1
- Step: 0.001
- Default: 0.999
- Purpose: Nucleus sampling parameter controlling diversity.
Top K
- Range: 0 to 500
- Step: 1
- Default: 128
- Purpose: Limits the number of highest probability tokens to consider.
Reasoning Budget Tokens
- Range: 1,024 to 64,000
- Step: 10
- Default: 1,024
- Purpose: Token budget allocated for reasoning and thinking (Claude extended thinking).
Multi-Environment Model Configuration
Configure different models for different environments:Model Selection Best Practices
- Cost optimization: Use smaller models (
amazon-nova-lite,claude-v3-haiku) for title generation - Performance: Use latest Claude models for best reasoning and accuracy
- Compliance: Restrict
globalAvailableModelsto organization-approved models only - Regional availability: Choose
bedrockRegionwith best model availability for your use case - Throughput: Enable global/cross-region inference for high-traffic deployments