AWS.Tools.BedrockRuntime.XML

<?xml version="1.0"?>
<doc>
    <assembly>
        <name>AWS.Tools.BedrockRuntime</name>
    </assembly>
    <members>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeCmdlet">
            <summary>
            Retrieve information about an asynchronous invocation.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeCmdlet.InvocationArn">
            <summary>
            <para>
            <para>The invocation's ARN.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is '*'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.GetAsyncInvokeResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.GetAsyncInvokeResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the InvocationArn parameter.
            The -PassThru parameter is deprecated, use -Select '^InvocationArn' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet">
            <summary>
            Lists asynchronous invocations.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.SortBy">
            <summary>
            <para>
            <para>How to sort the response.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.SortOrder">
            <summary>
            <para>
            <para>The sorting order for the response.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.StatusEqual">
            <summary>
            <para>
            <para>Filter invocations by status.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.SubmitTimeAfter">
            <summary>
            <para>
            <para>Include invocations submitted after this time.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.SubmitTimeBefore">
            <summary>
            <para>
            <para>Include invocations submitted before this time.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.MaxResult">
            <summary>
            <para>
            <para>The maximum number of invocations to return in one page of results.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.NextToken">
            <summary>
            <para>
            <para>Specify the pagination token from a previous request to retrieve the next page of
            results.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.GetBDRRAsyncInvokeListCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is 'AsyncInvokeSummaries'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ListAsyncInvokesResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.ListAsyncInvokesResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet">
            <summary>
            Sends messages to the specified Amazon Bedrock model. <c>Converse</c> provides a consistent
            interface that works with all models that support messages. This allows you to write
            code once and use it with different models. If a model has unique inference parameters,
            you can also pass those unique parameters to the model.
             
              
            <para>
            Amazon Bedrock doesn't store any text, images, or documents that you provide as content.
            The data is only used to generate the response.
            </para><para>
            You can submit a prompt by including it in the <c>messages</c> field, specifying the
            <c>modelId</c> of a foundation model or inference profile to run inference on it,
            and including any other fields that are relevant to your use case.
            </para><para>
            You can also submit a prompt from Prompt management by specifying the ARN of the prompt
            version and including a map of variables to values in the <c>promptVariables</c> field.
            You can append more messages to the prompt by using the <c>messages</c> field. If
            you use a prompt from Prompt management, you can't include the following fields in
            the request: <c>additionalModelRequestFields</c>, <c>inferenceConfig</c>, <c>system</c>,
            or <c>toolConfig</c>. Instead, these fields must be defined through Prompt management.
            For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management-use.html">Use
            a prompt from Prompt management</a>.
            </para><para>
            For information about the Converse API, see <i>Use the Converse API</i> in the <i>Amazon
            Bedrock User Guide</i>. To use a guardrail, see <i>Use a guardrail with the Converse
            API</i> in the <i>Amazon Bedrock User Guide</i>. To use a tool with a model, see <i>Tool
            use (Function calling)</i> in the <i>Amazon Bedrock User Guide</i></para><para>
            For example code, see <i>Converse API examples</i> in the <i>Amazon Bedrock User Guide</i>.
             
            </para><para>
            This operation requires permission for the <c>bedrock:InvokeModel</c> action.
            </para><important><para>
            To deny all inference access to resources that you specify in the modelId field, you
            need to deny access to the <c>bedrock:InvokeModel</c> and <c>bedrock:InvokeModelWithResponseStream</c>
            actions. Doing this also denies access to the resource through the base inference
            actions (<a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_InvokeModel.html">InvokeModel</a>
            and <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_InvokeModelWithResponseStream.html">InvokeModelWithResponseStream</a>).
            For more information see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/security_iam_id-based-policy-examples.html#security_iam_id-based-policy-examples-deny-inference">Deny
            access for inference on specific models</a>.
            </para></important><para>
            For troubleshooting some of the common errors you might encounter when using the <c>Converse</c>
            API, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/troubleshooting-api-error-codes.html">Troubleshooting
            Amazon Bedrock API Error Codes</a> in the Amazon Bedrock User Guide
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.AdditionalModelRequestField">
            <summary>
            <para>
            <para>Additional inference parameters that the model supports, beyond the base set of inference
            parameters that <c>Converse</c> and <c>ConverseStream</c> support in the <c>inferenceConfig</c>
            field. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Model
            parameters</a>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.AdditionalModelResponseFieldPath">
            <summary>
            <para>
            <para>Additional model parameters field paths to return in the response. <c>Converse</c>
            and <c>ConverseStream</c> return the requested fields as a JSON Pointer object in
            the <c>additionalModelResponseFields</c> field. The following is example JSON for
            <c>additionalModelResponseFieldPaths</c>.</para><para><c>[ "/stop_sequence" ]</c></para><para>For information about the JSON Pointer syntax, see the <a href="https://datatracker.ietf.org/doc/html/rfc6901">Internet
            Engineering Task Force (IETF)</a> documentation.</para><para><c>Converse</c> and <c>ConverseStream</c> reject an empty JSON Pointer or incorrectly
            structured JSON Pointer with a <c>400</c> error code. if the JSON Pointer is valid,
            but the requested field is not in the model response, it is ignored by <c>Converse</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolChoice_Any">
            <summary>
            <para>
            <para>The model must request at least one tool (no text is generated).</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolChoice_Auto">
            <summary>
            <para>
            <para>(Default). The Model automatically decides if a tool should be called or whether to
            generate text instead. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_GuardrailIdentifier">
            <summary>
            <para>
            <para>The identifier for the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_GuardrailVersion">
            <summary>
            <para>
            <para>The version of the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.PerformanceConfig_Latency">
            <summary>
            <para>
            <para>To use a latency-optimized version of the model, set to <c>optimized</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_MaxToken">
            <summary>
            <para>
            <para>The maximum number of tokens to allow in the generated response. The default value
            is the maximum allowed value for the model that you are using. For more information,
            see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Message">
            <summary>
            <para>
            <para>The messages that you want to send to the model.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ModelId">
            <summary>
            <para>
            <para>Specifies the model or throughput with which to run inference, or the prompt resource
            to use in inference. The value depends on the resource that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs
            for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon
            Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For
            a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported
            Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For
            more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run
            inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify
            the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use
            a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>To include a prompt that was defined in <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management.html">Prompt
            management</a>, specify the ARN of the prompt version to use.</para></li></ul><para>The Converse API doesn't support <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported
            models</a>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Tool_Name">
            <summary>
            <para>
            <para>The name of the tool that the model must request. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.PromptVariable">
            <summary>
            <para>
            <para>Contains a map of variables in a prompt from Prompt management to objects containing
            the values to fill in for them when running model invocation. This field is ignored
            if you don't specify a prompt resource in the <c>modelId</c> field.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.RequestMetadata">
            <summary>
            <para>
            <para>Key-value pairs that you can use to filter invocation logs.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_StopSequence">
            <summary>
            <para>
            <para>A list of stop sequences. A stop sequence is a sequence of characters that causes
            the model to stop generating the response. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.System">
            <summary>
            <para>
            <para>A prompt that provides instructions or context to the model about the task it should
            perform, or the persona it should adopt during the conversation.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_Temperature">
            <summary>
            <para>
            <para>The likelihood of the model selecting higher-probability options while generating
            a response. A lower value makes the model more likely to choose higher-probability
            options, while a higher value makes the model more likely to choose lower-probability
            options.</para><para>The default value is the default value for the model that you are using. For more
            information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.ToolConfig_Tool">
            <summary>
            <para>
            <para>An array of tools that you want to pass to a model.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.InferenceConfig_TopP">
            <summary>
            <para>
            <para>The percentage of most-likely candidates that the model considers for the next token.
            For example, if you choose a value of 0.8 for <c>topP</c>, the model selects from
            the top 80% of the probability distribution of tokens that could be next in the sequence.</para><para>The default value is the default value for the model that you are using. For more
            information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.GuardrailConfig_Trace">
            <summary>
            <para>
            <para>The trace behavior for the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is '*'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ConverseResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.ConverseResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the ModelId parameter.
            The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet">
            <summary>
            Sends messages to the specified Amazon Bedrock model and returns the response in a
            stream. <c>ConverseStream</c> provides a consistent API that works with all Amazon
            Bedrock models that support messages. This allows you to write code once and use it
            with different models. Should a model have unique inference parameters, you can also
            pass those unique parameters to the model.
             
              
            <para>
            To find out if a model supports streaming, call <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_GetFoundationModel.html">GetFoundationModel</a>
            and check the <c>responseStreamingSupported</c> field in the response.
            </para><note><para>
            The CLI doesn't support streaming operations in Amazon Bedrock, including <c>ConverseStream</c>.
            </para></note><para>
            Amazon Bedrock doesn't store any text, images, or documents that you provide as content.
            The data is only used to generate the response.
            </para><para>
            You can submit a prompt by including it in the <c>messages</c> field, specifying the
            <c>modelId</c> of a foundation model or inference profile to run inference on it,
            and including any other fields that are relevant to your use case.
            </para><para>
            You can also submit a prompt from Prompt management by specifying the ARN of the prompt
            version and including a map of variables to values in the <c>promptVariables</c> field.
            You can append more messages to the prompt by using the <c>messages</c> field. If
            you use a prompt from Prompt management, you can't include the following fields in
            the request: <c>additionalModelRequestFields</c>, <c>inferenceConfig</c>, <c>system</c>,
            or <c>toolConfig</c>. Instead, these fields must be defined through Prompt management.
            For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management-use.html">Use
            a prompt from Prompt management</a>.
            </para><para>
            For information about the Converse API, see <i>Use the Converse API</i> in the <i>Amazon
            Bedrock User Guide</i>. To use a guardrail, see <i>Use a guardrail with the Converse
            API</i> in the <i>Amazon Bedrock User Guide</i>. To use a tool with a model, see <i>Tool
            use (Function calling)</i> in the <i>Amazon Bedrock User Guide</i></para><para>
            For example code, see <i>Conversation streaming example</i> in the <i>Amazon Bedrock
            User Guide</i>.
            </para><para>
            This operation requires permission for the <c>bedrock:InvokeModelWithResponseStream</c>
            action.
            </para><important><para>
            To deny all inference access to resources that you specify in the modelId field, you
            need to deny access to the <c>bedrock:InvokeModel</c> and <c>bedrock:InvokeModelWithResponseStream</c>
            actions. Doing this also denies access to the resource through the base inference
            actions (<a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_InvokeModel.html">InvokeModel</a>
            and <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_InvokeModelWithResponseStream.html">InvokeModelWithResponseStream</a>).
            For more information see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/security_iam_id-based-policy-examples.html#security_iam_id-based-policy-examples-deny-inference">Deny
            access for inference on specific models</a>.
            </para></important><para>
            For troubleshooting some of the common errors you might encounter when using the <c>ConverseStream</c>
            API, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/troubleshooting-api-error-codes.html">Troubleshooting
            Amazon Bedrock API Error Codes</a> in the Amazon Bedrock User Guide
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.AdditionalModelRequestField">
            <summary>
            <para>
            <para>Additional inference parameters that the model supports, beyond the base set of inference
            parameters that <c>Converse</c> and <c>ConverseStream</c> support in the <c>inferenceConfig</c>
            field. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Model
            parameters</a>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.AdditionalModelResponseFieldPath">
            <summary>
            <para>
            <para>Additional model parameters field paths to return in the response. <c>Converse</c>
            and <c>ConverseStream</c> return the requested fields as a JSON Pointer object in
            the <c>additionalModelResponseFields</c> field. The following is example JSON for
            <c>additionalModelResponseFieldPaths</c>.</para><para><c>[ "/stop_sequence" ]</c></para><para>For information about the JSON Pointer syntax, see the <a href="https://datatracker.ietf.org/doc/html/rfc6901">Internet
            Engineering Task Force (IETF)</a> documentation.</para><para><c>Converse</c> and <c>ConverseStream</c> reject an empty JSON Pointer or incorrectly
            structured JSON Pointer with a <c>400</c> error code. if the JSON Pointer is valid,
            but the requested field is not in the model response, it is ignored by <c>Converse</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolChoice_Any">
            <summary>
            <para>
            <para>The model must request at least one tool (no text is generated).</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolChoice_Auto">
            <summary>
            <para>
            <para>(Default). The Model automatically decides if a tool should be called or whether to
            generate text instead. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_GuardrailIdentifier">
            <summary>
            <para>
            <para>The identifier for the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_GuardrailVersion">
            <summary>
            <para>
            <para>The version of the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.PerformanceConfig_Latency">
            <summary>
            <para>
            <para>To use a latency-optimized version of the model, set to <c>optimized</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_MaxToken">
            <summary>
            <para>
            <para>The maximum number of tokens to allow in the generated response. The default value
            is the maximum allowed value for the model that you are using. For more information,
            see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Message">
            <summary>
            <para>
            <para>The messages that you want to send to the model.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ModelId">
            <summary>
            <para>
            <para>Specifies the model or throughput with which to run inference, or the prompt resource
            to use in inference. The value depends on the resource that you use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs
            for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon
            Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For
            a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported
            Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For
            more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run
            inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify
            the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use
            a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>To include a prompt that was defined in <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-management.html">Prompt
            management</a>, specify the ARN of the prompt version to use.</para></li></ul><para>The Converse API doesn't support <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported
            models</a>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Tool_Name">
            <summary>
            <para>
            <para>The name of the tool that the model must request. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.PromptVariable">
            <summary>
            <para>
            <para>Contains a map of variables in a prompt from Prompt management to objects containing
            the values to fill in for them when running model invocation. This field is ignored
            if you don't specify a prompt resource in the <c>modelId</c> field.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.RequestMetadata">
            <summary>
            <para>
            <para>Key-value pairs that you can use to filter invocation logs.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_StopSequence">
            <summary>
            <para>
            <para>A list of stop sequences. A stop sequence is a sequence of characters that causes
            the model to stop generating the response. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_StreamProcessingMode">
            <summary>
            <para>
            <para>The processing mode. </para><para>The processing mode. For more information, see <i>Configure streaming response behavior</i>
            in the <i>Amazon Bedrock User Guide</i>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.System">
            <summary>
            <para>
            <para>A prompt that provides instructions or context to the model about the task it should
            perform, or the persona it should adopt during the conversation.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_Temperature">
            <summary>
            <para>
            <para>The likelihood of the model selecting higher-probability options while generating
            a response. A lower value makes the model more likely to choose higher-probability
            options, while a higher value makes the model more likely to choose lower-probability
            options.</para><para>The default value is the default value for the model that you are using. For more
            information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.ToolConfig_Tool">
            <summary>
            <para>
            <para>An array of tools that you want to pass to a model.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.InferenceConfig_TopP">
            <summary>
            <para>
            <para>The percentage of most-likely candidates that the model considers for the next token.
            For example, if you choose a value of 0.8 for <c>topP</c>, the model selects from
            the top 80% of the probability distribution of tokens that could be next in the sequence.</para><para>The default value is the default value for the model that you are using. For more
            information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters for foundation models</a>. </para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.GuardrailConfig_Trace">
            <summary>
            <para>
            <para>The trace behavior for the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is 'Stream'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ConverseStreamResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.ConverseStreamResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the ModelId parameter.
            The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRConverseStreamCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet">
            <summary>
            The action to apply a guardrail.
             
              
            <para>
            For troubleshooting some of the common errors you might encounter when using the <c>ApplyGuardrail</c>
            API, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/troubleshooting-api-error-codes.html">Troubleshooting
            Amazon Bedrock API Error Codes</a> in the Amazon Bedrock User Guide
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Content">
            <summary>
            <para>
            <para>The content details used in the request to apply the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.GuardrailIdentifier">
            <summary>
            <para>
            <para>The guardrail identifier used in the request to apply the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.GuardrailVersion">
            <summary>
            <para>
            <para>The guardrail version used in the request to apply the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.OutputScope">
            <summary>
            <para>
            <para>Specifies the scope of the output that you get in the response. Set to <c>FULL</c>
            to return the entire output, including any detected and non-detected entries in the
            response for enhanced debugging.</para><para>Note that the full output scope doesn't apply to word filters or regex in sensitive
            information filters. It does apply to all other filtering policies, including sensitive
            information with filters that can detect personally identifiable information (PII).</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Source">
            <summary>
            <para>
            <para>The source of data used in the request to apply the guardrail.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is '*'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.ApplyGuardrailResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.ApplyGuardrailResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the GuardrailIdentifier parameter.
            The -PassThru parameter is deprecated, use -Select '^GuardrailIdentifier' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRGuardrailCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet">
            <summary>
            Invokes the specified Amazon Bedrock model to run inference using the prompt and inference
            parameters provided in the request body. You use model inference to generate text,
            images, and embeddings.
             
              
            <para>
            For example code, see <i>Invoke model code examples</i> in the <i>Amazon Bedrock User
            Guide</i>.
            </para><para>
            This operation requires permission for the <c>bedrock:InvokeModel</c> action.
            </para><important><para>
            To deny all inference access to resources that you specify in the modelId field, you
            need to deny access to the <c>bedrock:InvokeModel</c> and <c>bedrock:InvokeModelWithResponseStream</c>
            actions. Doing this also denies access to the resource through the Converse API actions
            (<a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_Converse.html">Converse</a>
            and <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_ConverseStream.html">ConverseStream</a>).
            For more information see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/security_iam_id-based-policy-examples.html#security_iam_id-based-policy-examples-deny-inference">Deny
            access for inference on specific models</a>.
            </para></important><para>
            For troubleshooting some of the common errors you might encounter when using the <c>InvokeModel</c>
            API, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/troubleshooting-api-error-codes.html">Troubleshooting
            Amazon Bedrock API Error Codes</a> in the Amazon Bedrock User Guide
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Accept">
            <summary>
            <para>
            <para>The desired MIME type of the inference body in the response. The default value is
            <c>application/json</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Body">
            <summary>
            <para>
            <para>The prompt and inference parameters in the format specified in the <c>contentType</c>
            in the header. You must provide the body in JSON format. To see the format and content
            of the request and response bodies for different models, refer to <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters</a>. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/api-methods-run.html">Run
            inference</a> in the Bedrock User Guide.</para>
            </para>
            <para>The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.</para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.ContentType">
            <summary>
            <para>
            <para>The MIME type of the input data in the request. You must specify <c>application/json</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.GuardrailIdentifier">
            <summary>
            <para>
            <para>The unique identifier of the guardrail that you want to use. If you don't provide
            a value, no guardrail is applied to the invocation.</para><para>An error will be thrown in the following situations.</para><ul><li><para>You don't provide a guardrail identifier but you specify the <c>amazon-bedrock-guardrailConfig</c>
            field in the request body.</para></li><li><para>You enable the guardrail but the <c>contentType</c> isn't <c>application/json</c>.</para></li><li><para>You provide a guardrail identifier, but <c>guardrailVersion</c> isn't specified.</para></li></ul>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.GuardrailVersion">
            <summary>
            <para>
            <para>The version number for the guardrail. The value can also be <c>DRAFT</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.ModelId">
            <summary>
            <para>
            <para>The unique identifier of the model to invoke to run inference.</para><para>The <c>modelId</c> to provide depends on the type of model or throughput that you
            use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs
            for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon
            Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For
            a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported
            Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For
            more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run
            inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify
            the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use
            a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported
            model</a>, specify the ARN of the imported model. You can get the model ARN from a
            successful call to <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_CreateModelImportJob.html">CreateModelImportJob</a>
            or from the Imported models page in the Amazon Bedrock console.</para></li></ul>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.PerformanceConfigLatency">
            <summary>
            <para>
            <para>Model performance settings for the request.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Trace">
            <summary>
            <para>
            <para>Specifies whether to enable or disable the Bedrock trace. If enabled, you can see
            the full Bedrock trace.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is '*'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.InvokeModelResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.InvokeModelResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the ModelId parameter.
            The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet">
            <summary>
            Invoke the specified Amazon Bedrock model to run inference using the prompt and inference
            parameters provided in the request body. The response is returned in a stream.
             
              
            <para>
            To see if a model supports streaming, call <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_GetFoundationModel.html">GetFoundationModel</a>
            and check the <c>responseStreamingSupported</c> field in the response.
            </para><note><para>
            The CLI doesn't support streaming operations in Amazon Bedrock, including <c>InvokeModelWithResponseStream</c>.
            </para></note><para>
            For example code, see <i>Invoke model with streaming code example</i> in the <i>Amazon
            Bedrock User Guide</i>.
            </para><para>
            This operation requires permissions to perform the <c>bedrock:InvokeModelWithResponseStream</c>
            action.
            </para><important><para>
            To deny all inference access to resources that you specify in the modelId field, you
            need to deny access to the <c>bedrock:InvokeModel</c> and <c>bedrock:InvokeModelWithResponseStream</c>
            actions. Doing this also denies access to the resource through the Converse API actions
            (<a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_Converse.html">Converse</a>
            and <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_ConverseStream.html">ConverseStream</a>).
            For more information see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/security_iam_id-based-policy-examples.html#security_iam_id-based-policy-examples-deny-inference">Deny
            access for inference on specific models</a>.
            </para></important><para>
            For troubleshooting some of the common errors you might encounter when using the <c>InvokeModelWithResponseStream</c>
            API, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/troubleshooting-api-error-codes.html">Troubleshooting
            Amazon Bedrock API Error Codes</a> in the Amazon Bedrock User Guide
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Accept">
            <summary>
            <para>
            <para>The desired MIME type of the inference body in the response. The default value is
            <c>application/json</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Body">
            <summary>
            <para>
            <para>The prompt and inference parameters in the format specified in the <c>contentType</c>
            in the header. You must provide the body in JSON format. To see the format and content
            of the request and response bodies for different models, refer to <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html">Inference
            parameters</a>. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/api-methods-run.html">Run
            inference</a> in the Bedrock User Guide.</para>
            </para>
            <para>The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.</para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.ContentType">
            <summary>
            <para>
            <para>The MIME type of the input data in the request. You must specify <c>application/json</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.GuardrailIdentifier">
            <summary>
            <para>
            <para>The unique identifier of the guardrail that you want to use. If you don't provide
            a value, no guardrail is applied to the invocation.</para><para>An error is thrown in the following situations.</para><ul><li><para>You don't provide a guardrail identifier but you specify the <c>amazon-bedrock-guardrailConfig</c>
            field in the request body.</para></li><li><para>You enable the guardrail but the <c>contentType</c> isn't <c>application/json</c>.</para></li><li><para>You provide a guardrail identifier, but <c>guardrailVersion</c> isn't specified.</para></li></ul>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.GuardrailVersion">
            <summary>
            <para>
            <para>The version number for the guardrail. The value can also be <c>DRAFT</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.ModelId">
            <summary>
            <para>
            <para>The unique identifier of the model to invoke to run inference.</para><para>The <c>modelId</c> to provide depends on the type of model or throughput that you
            use:</para><ul><li><para>If you use a base model, specify the model ID or its ARN. For a list of model IDs
            for base models, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-ids.html#model-ids-arns">Amazon
            Bedrock base model IDs (on-demand throughput)</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an inference profile, specify the inference profile ID or its ARN. For
            a list of inference profile IDs, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/cross-region-inference-support.html">Supported
            Regions and models for cross-region inference</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a provisioned model, specify the ARN of the Provisioned Throughput. For
            more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/prov-thru-use.html">Run
            inference using a Provisioned Throughput</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use a custom model, first purchase Provisioned Throughput for it. Then specify
            the ARN of the resulting provisioned model. For more information, see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-use.html">Use
            a custom model in Amazon Bedrock</a> in the Amazon Bedrock User Guide.</para></li><li><para>If you use an <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/model-customization-import-model.html">imported
            model</a>, specify the ARN of the imported model. You can get the model ARN from a
            successful call to <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_CreateModelImportJob.html">CreateModelImportJob</a>
            or from the Imported models page in the Amazon Bedrock console.</para></li></ul>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.PerformanceConfigLatency">
            <summary>
            <para>
            <para>Model performance settings for the request.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Trace">
            <summary>
            <para>
            <para>Specifies whether to enable or disable the Bedrock trace. If enabled, you can see
            the full Bedrock trace.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is '*'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.InvokeModelWithResponseStreamResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.InvokeModelWithResponseStreamResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the ModelId parameter.
            The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.InvokeBDRRModelWithResponseStreamCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
        <member name="T:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet">
            <summary>
            Starts an asynchronous invocation.
             
              
            <para>
            This operation requires permission for the <c>bedrock:InvokeModel</c> action.
            </para><important><para>
            To deny all inference access to resources that you specify in the modelId field, you
            need to deny access to the <c>bedrock:InvokeModel</c> and <c>bedrock:InvokeModelWithResponseStream</c>
            actions. Doing this also denies access to the resource through the Converse API actions
            (<a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_Converse.html">Converse</a>
            and <a href="https://docs.aws.amazon.com/bedrock/latest/APIReference/API_runtime_ConverseStream.html">ConverseStream</a>).
            For more information see <a href="https://docs.aws.amazon.com/bedrock/latest/userguide/security_iam_id-based-policy-examples.html#security_iam_id-based-policy-examples-deny-inference">Deny
            access for inference on specific models</a>.
            </para></important>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.S3OutputDataConfig_BucketOwner">
            <summary>
            <para>
            <para>If the bucket belongs to another AWS account, specify that account's ID.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.ClientRequestToken">
            <summary>
            <para>
            <para>Specify idempotency token to ensure that requests are not duplicated.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.S3OutputDataConfig_KmsKeyId">
            <summary>
            <para>
            <para>A KMS encryption key ID.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.ModelId">
            <summary>
            <para>
            <para>The model to invoke.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.ModelInput">
            <summary>
            <para>
            <para>Input to send to the model.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.S3OutputDataConfig_S3Uri">
            <summary>
            <para>
            <para>An object URI starting with <c>s3://</c>.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.Tag">
            <summary>
            <para>
            <para>Tags to apply to the invocation.</para>
            </para>
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.Select">
            <summary>
            Use the -Select parameter to control the cmdlet output. The default value is 'InvocationArn'.
            Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.BedrockRuntime.Model.StartAsyncInvokeResponse).
            Specifying the name of a property of type Amazon.BedrockRuntime.Model.StartAsyncInvokeResponse will result in that property being returned.
            Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.PassThru">
            <summary>
            Changes the cmdlet behavior to return the value passed to the ModelId parameter.
            The -PassThru parameter is deprecated, use -Select '^ModelId' instead. This parameter will be removed in a future version.
            </summary>
        </member>
        <member name="P:Amazon.PowerShell.Cmdlets.BDRR.StartBDRRAsyncInvokeCmdlet.Force">
            <summary>
            This parameter overrides confirmation prompts to force
            the cmdlet to continue its operation. This parameter should always
            be used with caution.
            </summary>
        </member>
    </members>
</doc>