AWS.Tools.BedrockRuntime.XML
<?xml version="1.0"?>
<doc> <assembly> <name>AWS.Tools.BedrockRuntime</name> </assembly> <members> <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet"> <summary> Sends messages to the specified Amazon Bedrock model. <c>Converse</c> provides a consistent interface that works with all models that support messages. This allows you to write code once and use it with different models. If a model has unique inference parameters, you can also pass those unique parameters to the model. <para> Amazon Bedrock doesn't store any text, images, or documents that you provide as content. The data is only used to generate the response. </para><para> You can submit a prompt by including it in the <c>messages</c> field, specifying the <c>modelId</c> of a foundation model or inference profile to run inference on it, and including any other fields that are relevant to your use case. </para><para> You can also submit a prompt from Prompt management by specifying the ARN of the prompt version and including a map of variables to values in the <c>promptVariables</c> field. You can append more messages to the prompt by using the <c>messages</c> field. If you use a prompt from Prompt management, you can't include the following fields in the request: <c>additionalModelRequestFields</c>, <c>inferenceConfig</c>, <c>system</c>, or <c>toolConfig</c>. Instead, these fields must be defined through Prompt management. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management-use.html">Use a prompt from Prompt management</a>. </para><para> For information about the Converse API, see <i>Use the Converse API</i> in the <i>Amazon Bedrock User Guide</i>. To use a guardrail, see <i>Use a guardrail with the Converse API</i> in the <i>Amazon Bedrock User Guide</i>. To use a tool with a model, see <i>Tool use (Function calling)</i> in the <i>Amazon Bedrock User Guide</i></para><para> For example code, see <i>Converse API examples</i> in the <i>Amazon Bedrock User Guide</i>. </para><para> This operation requires permission for the <c>bedrock:InvokeModel</c> action. </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.AdditionalModelRequestField"> <summary> <para> <para>Additional inference parameters that the model supports, beyond the base set of inference parameters that <c>Converse</c> and <c>ConverseStream</c> support in the <c>inferenceConfig</c> field. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Model parameters</a>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.AdditionalModelResponseFieldPath"> <summary> <para> <para>Additional model parameters field paths to return in the response. <c>Converse</c> and <c>ConverseStream</c> return the requested fields as a JSON Pointer object in the <c>additionalModelResponseFields</c> field. The following is example JSON for <c>additionalModelResponseFieldPaths</c>.</para><para><c>[ "/stop_sequence" ]</c></para><para>For information about the JSON Pointer syntax, see the <a href="https://datatracker.ietf.org/doc/html/rfc6901">Internet Engineering Task Force (IETF)</a> documentation.</para><para><c>Converse</c> and <c>ConverseStream</c> reject an empty JSON Pointer or incorrectly structured JSON Pointer with a <c>400</c> error code. if the JSON Pointer is valid, but the requested field is not in the model response, it is ignored by <c>Converse</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolChoice_Any"> <summary> <para> <para>The model must request at least one tool (no text is generated).</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolChoice_Auto"> <summary> <para> <para>(Default). The Model automatically decides if a tool should be called or whether to generate text instead. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_GuardrailIdentifier"> <summary> <para> <para>The identifier for the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_GuardrailVersion"> <summary> <para> <para>The version of the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_MaxToken"> <summary> <para> <para>The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Message"> <summary> <para> <para>The messages that you want to send to the model.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ModelId"> <summary> <para> <para>Specifies the model or throughput with which to run inference, or the prompt resource to use in inference. The value depends on the resource that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>To include a prompt that was defined in Prompt management, specify the ARN of the prompt version to use.</para></li></ul><para>The Converse API doesn't support <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported models</a>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Tool_Name"> <summary> <para> <para>The name of the tool that the model must request. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.PromptVariable"> <summary> <para> <para>Contains a map of variables in a prompt from Prompt management to objects containing the values to fill in for them when running model invocation. This field is ignored if you don't specify a prompt resource in the <c>modelId</c> field.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_StopSequence"> <summary> <para> <para>A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.System"> <summary> <para> <para>A prompt that provides instructions or context to the model about the task it should perform, or the persona it should adopt during the conversation.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_Temperature"> <summary> <para> <para>The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options, while a higher value makes the model more likely to choose lower-probability options.</para><para>The default value is the default value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolConfig_Tool"> <summary> <para> <para>An array of tools that you want to pass to a model.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_TopP"> <summary> <para> <para>The percentage of most-likely candidates that the model considers for the next token. For example, if you choose a value of 0.8 for <c>topP</c>, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence.</para><para>The default value is the default value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_Trace"> <summary> <para> <para>The trace behavior for the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Select"> <summary> Use the -Select parameter to control the cmdlet output. The default value is '*'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ConverseResponse). Specifying the name of a property of type Amazon.BedrockRuntime.Model.ConverseResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.PassThru"> <summary> Changes the cmdlet behavior to return the value passed to the ModelId parameter. The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Force"> <summary> This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution. </summary> </member> <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet"> <summary> Sends messages to the specified Amazon Bedrock model and returns the response in a stream. <c>ConverseStream</c> provides a consistent API that works with all Amazon Bedrock models that support messages. This allows you to write code once and use it with different models. Should a model have unique inference parameters, you can also pass those unique parameters to the model. <para> To find out if a model supports streaming, call <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_GetFoundationModel.html">GetFoundationModel</a> and check the <c>responseStreamingSupported</c> field in the response. </para><note><para> The CLI doesn't support streaming operations in Amazon Bedrock, including <c>ConverseStream</c>. </para></note><para> Amazon Bedrock doesn't store any text, images, or documents that you provide as content. The data is only used to generate the response. </para><para> You can submit a prompt by including it in the <c>messages</c> field, specifying the <c>modelId</c> of a foundation model or inference profile to run inference on it, and including any other fields that are relevant to your use case. </para><para> You can also submit a prompt from Prompt management by specifying the ARN of the prompt version and including a map of variables to values in the <c>promptVariables</c> field. You can append more messages to the prompt by using the <c>messages</c> field. If you use a prompt from Prompt management, you can't include the following fields in the request: <c>additionalModelRequestFields</c>, <c>inferenceConfig</c>, <c>system</c>, or <c>toolConfig</c>. Instead, these fields must be defined through Prompt management. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management-use.html">Use a prompt from Prompt management</a>. </para><para> For information about the Converse API, see <i>Use the Converse API</i> in the <i>Amazon Bedrock User Guide</i>. To use a guardrail, see <i>Use a guardrail with the Converse API</i> in the <i>Amazon Bedrock User Guide</i>. To use a tool with a model, see <i>Tool use (Function calling)</i> in the <i>Amazon Bedrock User Guide</i></para><para> For example code, see <i>Conversation streaming example</i> in the <i>Amazon Bedrock User Guide</i>. </para><para> This operation requires permission for the <c>bedrock:InvokeModelWithResponseStream</c> action. </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.AdditionalModelRequestField"> <summary> <para> <para>Additional inference parameters that the model supports, beyond the base set of inference parameters that <c>Converse</c> and <c>ConverseStream</c> support in the <c>inferenceConfig</c> field. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Model parameters</a>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.AdditionalModelResponseFieldPath"> <summary> <para> <para>Additional model parameters field paths to return in the response. <c>Converse</c> and <c>ConverseStream</c> return the requested fields as a JSON Pointer object in the <c>additionalModelResponseFields</c> field. The following is example JSON for <c>additionalModelResponseFieldPaths</c>.</para><para><c>[ "/stop_sequence" ]</c></para><para>For information about the JSON Pointer syntax, see the <a href="https://datatracker.ietf.org/doc/html/rfc6901">Internet Engineering Task Force (IETF)</a> documentation.</para><para><c>Converse</c> and <c>ConverseStream</c> reject an empty JSON Pointer or incorrectly structured JSON Pointer with a <c>400</c> error code. if the JSON Pointer is valid, but the requested field is not in the model response, it is ignored by <c>Converse</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolChoice_Any"> <summary> <para> <para>The model must request at least one tool (no text is generated).</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolChoice_Auto"> <summary> <para> <para>(Default). The Model automatically decides if a tool should be called or whether to generate text instead. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_GuardrailIdentifier"> <summary> <para> <para>The identifier for the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_GuardrailVersion"> <summary> <para> <para>The version of the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_MaxToken"> <summary> <para> <para>The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Message"> <summary> <para> <para>The messages that you want to send to the model.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ModelId"> <summary> <para> <para>Specifies the model or throughput with which to run inference, or the prompt resource to use in inference. The value depends on the resource that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>To include a prompt that was defined in Prompt management, specify the ARN of the prompt version to use.</para></li></ul><para>The Converse API doesn't support <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported models</a>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Tool_Name"> <summary> <para> <para>The name of the tool that the model must request. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.PromptVariable"> <summary> <para> <para>Contains a map of variables in a prompt from Prompt management to objects containing the values to fill in for them when running model invocation. This field is ignored if you don't specify a prompt resource in the <c>modelId</c> field.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_StopSequence"> <summary> <para> <para>A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_StreamProcessingMode"> <summary> <para> <para>The processing mode. </para><para>The processing mode. For more information, see <i>Configure streaming response behavior</i> in the <i>Amazon Bedrock User Guide</i>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.System"> <summary> <para> <para>A prompt that provides instructions or context to the model about the task it should perform, or the persona it should adopt during the conversation.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_Temperature"> <summary> <para> <para>The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options, while a higher value makes the model more likely to choose lower-probability options.</para><para>The default value is the default value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolConfig_Tool"> <summary> <para> <para>An array of tools that you want to pass to a model.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_TopP"> <summary> <para> <para>The percentage of most-likely candidates that the model considers for the next token. For example, if you choose a value of 0.8 for <c>topP</c>, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence.</para><para>The default value is the default value for the model that you are using. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters for foundation models</a>. </para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_Trace"> <summary> <para> <para>The trace behavior for the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Select"> <summary> Use the -Select parameter to control the cmdlet output. The default value is 'Stream'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ConverseStreamResponse). Specifying the name of a property of type Amazon.BedrockRuntime.Model.ConverseStreamResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.PassThru"> <summary> Changes the cmdlet behavior to return the value passed to the ModelId parameter. The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Force"> <summary> This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution. </summary> </member> <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet"> <summary> The action to apply a guardrail. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Content"> <summary> <para> <para>The content details used in the request to apply the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.GuardrailIdentifier"> <summary> <para> <para>The guardrail identifier used in the request to apply the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.GuardrailVersion"> <summary> <para> <para>The guardrail version used in the request to apply the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Source"> <summary> <para> <para>The source of data used in the request to apply the guardrail.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Select"> <summary> Use the -Select parameter to control the cmdlet output. The default value is '*'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ApplyGuardrailResponse). Specifying the name of a property of type Amazon.BedrockRuntime.Model.ApplyGuardrailResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.PassThru"> <summary> Changes the cmdlet behavior to return the value passed to the GuardrailIdentifier parameter. The -PassThru parameter is deprecated, use -Select '^GuardrailIdentifier' instead. This parameter will be removed in a future version. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Force"> <summary> This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution. </summary> </member> <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet"> <summary> Invokes the specified Amazon Bedrock model to run inference using the prompt and inference parameters provided in the request body. You use model inference to generate text, images, and embeddings. <para> For example code, see <i>Invoke model code examples</i> in the <i>Amazon Bedrock User Guide</i>. </para><para> This operation requires permission for the <c>bedrock:InvokeModel</c> action. </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Accept"> <summary> <para> <para>The desired MIME type of the inference body in the response. The default value is <c>application/json</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Body"> <summary> <para> <para>The prompt and inference parameters in the format specified in the <c>contentType</c> in the header. You must provide the body in JSON format. To see the format and content of the request and response bodies for different models, refer to <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters</a>. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/api-methods-run.html">Run inference</a> in the Bedrock User Guide.</para> </para> <para>The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.</para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.ContentType"> <summary> <para> <para>The MIME type of the input data in the request. You must specify <c>application/json</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.GuardrailIdentifier"> <summary> <para> <para>The unique identifier of the guardrail that you want to use. If you don't provide a value, no guardrail is applied to the invocation.</para><para>An error will be thrown in the following situations.</para><ul><li><para>You don't provide a guardrail identifier but you specify the <c>amazon-bedrock-guardrailConfig</c> field in the request body.</para></li><li><para>You enable the guardrail but the <c>contentType</c> isn't <c>application/json</c>.</para></li><li><para>You provide a guardrail identifier, but <c>guardrailVersion</c> isn't specified.</para></li></ul> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.GuardrailVersion"> <summary> <para> <para>The version number for the guardrail. The value can also be <c>DRAFT</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.ModelId"> <summary> <para> <para>The unique identifier of the model to invoke to run inference.</para><para>The <c>modelId</c> to provide depends on the type of model that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported model</a>, specify the ARN of the imported model. You can get the model ARN from a successful call to <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_CreateModelImportJob.html">CreateModelImportJob</a> or from the Imported models page in the Amazon Bedrock console.</para></li></ul> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Trace"> <summary> <para> <para>Specifies whether to enable or disable the Bedrock trace. If enabled, you can see the full Bedrock trace.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Select"> <summary> Use the -Select parameter to control the cmdlet output. The default value is '*'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.InvokeModelResponse). Specifying the name of a property of type Amazon.BedrockRuntime.Model.InvokeModelResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.PassThru"> <summary> Changes the cmdlet behavior to return the value passed to the ModelId parameter. The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Force"> <summary> This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution. </summary> </member> <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet"> <summary> Invoke the specified Amazon Bedrock model to run inference using the prompt and inference parameters provided in the request body. The response is returned in a stream. <para> To see if a model supports streaming, call <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_GetFoundationModel.html">GetFoundationModel</a> and check the <c>responseStreamingSupported</c> field in the response. </para><note><para> The CLI doesn't support streaming operations in Amazon Bedrock, including <c>InvokeModelWithResponseStream</c>. </para></note><para> For example code, see <i>Invoke model with streaming code example</i> in the <i>Amazon Bedrock User Guide</i>. </para><para> This operation requires permissions to perform the <c>bedrock:InvokeModelWithResponseStream</c> action. </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Accept"> <summary> <para> <para>The desired MIME type of the inference body in the response. The default value is <c>application/json</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Body"> <summary> <para> <para>The prompt and inference parameters in the format specified in the <c>contentType</c> in the header. You must provide the body in JSON format. To see the format and content of the request and response bodies for different models, refer to <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference parameters</a>. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/api-methods-run.html">Run inference</a> in the Bedrock User Guide.</para> </para> <para>The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.</para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.ContentType"> <summary> <para> <para>The MIME type of the input data in the request. You must specify <c>application/json</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.GuardrailIdentifier"> <summary> <para> <para>The unique identifier of the guardrail that you want to use. If you don't provide a value, no guardrail is applied to the invocation.</para><para>An error is thrown in the following situations.</para><ul><li><para>You don't provide a guardrail identifier but you specify the <c>amazon-bedrock-guardrailConfig</c> field in the request body.</para></li><li><para>You enable the guardrail but the <c>contentType</c> isn't <c>application/json</c>.</para></li><li><para>You provide a guardrail identifier, but <c>guardrailVersion</c> isn't specified.</para></li></ul> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.GuardrailVersion"> <summary> <para> <para>The version number for the guardrail. The value can also be <c>DRAFT</c>.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.ModelId"> <summary> <para> <para>The unique identifier of the model to invoke to run inference.</para><para>The <c>modelId</c> to provide depends on the type of model that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported model</a>, specify the ARN of the imported model. You can get the model ARN from a successful call to <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_CreateModelImportJob.html">CreateModelImportJob</a> or from the Imported models page in the Amazon Bedrock console.</para></li></ul> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Trace"> <summary> <para> <para>Specifies whether to enable or disable the Bedrock trace. If enabled, you can see the full Bedrock trace.</para> </para> </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Select"> <summary> Use the -Select parameter to control the cmdlet output. The default value is '*'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.InvokeModelWithResponseStreamResponse). Specifying the name of a property of type Amazon.BedrockRuntime.Model.InvokeModelWithResponseStreamResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.PassThru"> <summary> Changes the cmdlet behavior to return the value passed to the ModelId parameter. The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version. </summary> </member> <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Force"> <summary> This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution. </summary> </member> </members> </doc> |