Bedrock / Client / get_inference_profile
get_inference_profile#
- Bedrock.Client.get_inference_profile(**kwargs)#
Gets information about an inference profile. For more information, see Increase throughput and resilience with cross-region inference in Amazon Bedrock. in the Amazon Bedrock User Guide.
See also: AWS API Documentation
Request Syntax
response = client.get_inference_profile( inferenceProfileIdentifier='string' )
- Parameters:
inferenceProfileIdentifier (string) –
[REQUIRED]
The ID or Amazon Resource Name (ARN) of the inference profile.
- Return type:
dict
- Returns:
Response Syntax
{ 'inferenceProfileName': 'string', 'description': 'string', 'createdAt': datetime(2015, 1, 1), 'updatedAt': datetime(2015, 1, 1), 'inferenceProfileArn': 'string', 'models': [ { 'modelArn': 'string' }, ], 'inferenceProfileId': 'string', 'status': 'ACTIVE', 'type': 'SYSTEM_DEFINED'|'APPLICATION' }
Response Structure
(dict) –
inferenceProfileName (string) –
The name of the inference profile.
description (string) –
The description of the inference profile.
createdAt (datetime) –
The time at which the inference profile was created.
updatedAt (datetime) –
The time at which the inference profile was last updated.
inferenceProfileArn (string) –
The Amazon Resource Name (ARN) of the inference profile.
models (list) –
A list of information about each model in the inference profile.
(dict) –
Contains information about a model.
modelArn (string) –
The Amazon Resource Name (ARN) of the model.
inferenceProfileId (string) –
The unique identifier of the inference profile.
status (string) –
The status of the inference profile.
ACTIVE
means that the inference profile is ready to be used.type (string) –
The type of the inference profile. The following types are possible:
SYSTEM_DEFINED
– The inference profile is defined by Amazon Bedrock. You can route inference requests across regions with these inference profiles.APPLICATION
– The inference profile was created by a user. This type of inference profile can track metrics and costs when invoking the model in it. The inference profile may route requests to one or multiple regions.
Exceptions