AWS Bedrock Object Structure
Integrating AWS Bedrock with your application allows you to seamlessly utilize multiple AI models hosted on AWS. This section details how to configure the AWS Bedrock endpoint for your needs.
Example Configuration
Note: AWS Bedrock endpoint supports all Shared Endpoint Settings, including
streamRate,titleModel,titleMethod,titlePrompt,titlePromptTemplate, andtitleEndpoint. The settings shown below are specific to Bedrock or have Bedrock-specific defaults.
titleModel
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| titleModel | String | Specifies the model to use for generating conversation titles. | Recommended: anthropic.claude-3-haiku-20240307-v1:0. Set to "current_model" to use the same model as the chat. |
Default: Not specified
Example:
streamRate
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| streamRate | Number | Sets the rate of processing each new token in milliseconds. | This can help stabilize processing of concurrent requests and provide smoother frontend stream rendering. |
Default: Not specified
Example:
availableRegions
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| availableRegions | Array | Specifies the AWS regions you want to make available for Bedrock. | If provided, users will see a dropdown to select the region. If not selected, the default region is used. |
Default: Not specified
Example:
models
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| models | Array of Strings | Specifies custom model IDs available for the Bedrock endpoint. | When provided, these models appear in the model selector for Bedrock. |
Default: Not specified (uses default Bedrock model list)
Example:
inferenceProfiles
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| inferenceProfiles | Object (Record) | Maps model IDs to inference profile ARNs for cross-region inference. Keys are model IDs and values are the inference profile ARN or an environment variable reference. | When a selected model matches a key, the corresponding ARN is used as the application inference profile. |
Default: Not specified
Example:
Notes:
- Inference profiles enable cross-region inference, allowing you to route requests to models in different AWS regions
- Values support environment variable interpolation with
${ENV_VAR}syntax - The model ID in the key must match the model selected by the user in the UI
- Use with the
modelsfield to make cross-region model IDs available in the model selector - For a complete guide on creating and managing inference profiles, see AWS Bedrock Inference Profiles
Combined Example:
guardrailConfig
Key:
| Key | Type | Description | Example |
|---|---|---|---|
| guardrailConfig | Object | Configuration for AWS Bedrock Guardrails to filter and moderate model inputs and outputs. | Optional. When configured, all Bedrock requests will be validated against the specified guardrail. |
Sub-keys:
| Key | Type | Description | Example |
|---|---|---|---|
| guardrailIdentifier | String | The unique identifier of the guardrail to apply. | Required when using guardrails. |
| guardrailVersion | String | The version of the guardrail to use. | Required when using guardrails. |
| trace | String | Controls guardrail trace output for debugging. Options: "enabled", "enabled_full", or "disabled". | Optional. Default: "disabled" |
Example:
Notes:
- Guardrails help ensure responsible AI usage by filtering harmful content, PII, and other sensitive information
- The
guardrailIdentifiercan be found in the AWS Bedrock console under Guardrails - Set
traceto"enabled"or"enabled_full"during development to see which guardrail policies are triggered - For production, set
traceto"disabled"to reduce response payload size
Notes
- The main configuration for AWS Bedrock is done through environment variables, additional forms of authentication are in development.
How is this guide?