|
31 | 31 | {"shape":"ResourceNotFoundException"}, |
32 | 32 | {"shape":"ThrottlingException"}, |
33 | 33 | {"shape":"InternalServerException"}, |
| 34 | + {"shape":"ServiceUnavailableException"}, |
34 | 35 | {"shape":"ValidationException"}, |
35 | 36 | {"shape":"ServiceQuotaExceededException"} |
36 | 37 | ], |
|
97 | 98 | {"shape":"ServiceUnavailableException"}, |
98 | 99 | {"shape":"ValidationException"} |
99 | 100 | ], |
100 | | - "documentation":"<p>Returns the token count for a given inference request. This operation helps you estimate token usage before sending requests to foundation models by returning the token count that would be used if the same input were sent to the model in an inference request.</p> <p>Token counting is model-specific because different models use different tokenization strategies. The token count returned by this operation will match the token count that would be charged if the same input were sent to the model in an <code>InvokeModel</code> or <code>Converse</code> request.</p> <p>You can use this operation to:</p> <ul> <li> <p>Estimate costs before sending inference requests.</p> </li> <li> <p>Optimize prompts to fit within token limits.</p> </li> <li> <p>Plan for token usage in your applications.</p> </li> </ul> <p>This operation accepts the same input formats as <code>InvokeModel</code> and <code>Converse</code>, allowing you to count tokens for both raw text inputs and structured conversation formats.</p> <p>The following operations are related to <code>CountTokens</code>:</p> <ul> <li> <p> <a href=\"https://docs.aws.amazon.com/bedrock/latest/API/API_runtime_InvokeModel.html\">InvokeModel</a> - Sends inference requests to foundation models</p> </li> <li> <p> <a href=\"https://docs.aws.amazon.com/bedrock/latest/API/API_runtime_Converse.html\">Converse</a> - Sends conversation-based inference requests to foundation models</p> </li> </ul>" |
| 101 | + "documentation":"<p>Returns the token count for a given inference request. This operation helps you estimate token usage before sending requests to foundation models by returning the token count that would be used if the same input were sent to the model in an inference request.</p> <p>Token counting is model-specific because different models use different tokenization strategies. The token count returned by this operation will match the token count that would be charged if the same input were sent to the model in an <code>InvokeModel</code> or <code>Converse</code> request.</p> <p>You can use this operation to:</p> <ul> <li> <p>Estimate costs before sending inference requests.</p> </li> <li> <p>Optimize prompts to fit within token limits.</p> </li> <li> <p>Plan for token usage in your applications.</p> </li> </ul> <p>This operation accepts the same input formats as <code>InvokeModel</code> and <code>Converse</code>, allowing you to count tokens for both raw text inputs and structured conversation formats.</p> <p>The following operations are related to <code>CountTokens</code>:</p> <ul> <li> <p> <a href=\"https://docs.aws.amazon.com/bedrock/latest/API/API_runtime_InvokeModel.html\">InvokeModel</a> - Sends inference requests to foundation models</p> </li> <li> <p> <a href=\"https://docs.aws.amazon.com/bedrock/latest/API/API_runtime_Converse.html\">Converse</a> - Sends conversation-based inference requests to foundation models</p> </li> </ul>", |
| 102 | + "readonly":true |
101 | 103 | }, |
102 | 104 | "GetAsyncInvoke":{ |
103 | 105 | "name":"GetAsyncInvoke", |
|
114 | 116 | {"shape":"InternalServerException"}, |
115 | 117 | {"shape":"ValidationException"} |
116 | 118 | ], |
117 | | - "documentation":"<p>Retrieve information about an asynchronous invocation.</p>" |
| 119 | + "documentation":"<p>Retrieve information about an asynchronous invocation.</p>", |
| 120 | + "readonly":true |
118 | 121 | }, |
119 | 122 | "InvokeModel":{ |
120 | 123 | "name":"InvokeModel", |
|
202 | 205 | {"shape":"InternalServerException"}, |
203 | 206 | {"shape":"ValidationException"} |
204 | 207 | ], |
205 | | - "documentation":"<p>Lists asynchronous invocations.</p>" |
| 208 | + "documentation":"<p>Lists asynchronous invocations.</p>", |
| 209 | + "readonly":true |
206 | 210 | }, |
207 | 211 | "StartAsyncInvoke":{ |
208 | 212 | "name":"StartAsyncInvoke", |
|
246 | 250 | }, |
247 | 251 | "AnyToolChoice":{ |
248 | 252 | "type":"structure", |
249 | | - "members":{ |
250 | | - }, |
| 253 | + "members":{}, |
251 | 254 | "documentation":"<p>The model must request at least one tool (no text is generated). For example, <code>{\"any\" : {}}</code>.</p>" |
252 | 255 | }, |
253 | 256 | "ApplyGuardrailRequest":{ |
|
436 | 439 | }, |
437 | 440 | "AutoToolChoice":{ |
438 | 441 | "type":"structure", |
439 | | - "members":{ |
440 | | - }, |
| 442 | + "members":{}, |
441 | 443 | "documentation":"<p>The Model automatically decides if a tool should be called or whether to generate text instead. For example, <code>{\"auto\" : {}}</code>.</p>" |
442 | 444 | }, |
443 | 445 | "AutomatedReasoningRuleIdentifier":{ |
|
1158 | 1160 | }, |
1159 | 1161 | "Document":{ |
1160 | 1162 | "type":"structure", |
1161 | | - "members":{ |
1162 | | - }, |
| 1163 | + "members":{}, |
1163 | 1164 | "document":true |
1164 | 1165 | }, |
1165 | 1166 | "DocumentBlock":{ |
|
1596 | 1597 | }, |
1597 | 1598 | "GuardrailAutomatedReasoningNoTranslationsFinding":{ |
1598 | 1599 | "type":"structure", |
1599 | | - "members":{ |
1600 | | - }, |
| 1600 | + "members":{}, |
1601 | 1601 | "documentation":"<p>Indicates that no relevant logical information could be extracted from the input for validation.</p>" |
1602 | 1602 | }, |
1603 | 1603 | "GuardrailAutomatedReasoningPoliciesProcessed":{ |
|
1706 | 1706 | }, |
1707 | 1707 | "GuardrailAutomatedReasoningTooComplexFinding":{ |
1708 | 1708 | "type":"structure", |
1709 | | - "members":{ |
1710 | | - }, |
| 1709 | + "members":{}, |
1711 | 1710 | "documentation":"<p>Indicates that the input exceeds the processing capacity due to the volume or complexity of the logical information.</p>" |
1712 | 1711 | }, |
1713 | 1712 | "GuardrailAutomatedReasoningTranslation":{ |
|
3177 | 3176 | }, |
3178 | 3177 | "ModelInputPayload":{ |
3179 | 3178 | "type":"structure", |
3180 | | - "members":{ |
3181 | | - }, |
| 3179 | + "members":{}, |
3182 | 3180 | "document":true, |
3183 | 3181 | "sensitive":true |
3184 | 3182 | }, |
|
3551 | 3549 | "max_tokens", |
3552 | 3550 | "stop_sequence", |
3553 | 3551 | "guardrail_intervened", |
3554 | | - "content_filtered" |
| 3552 | + "content_filtered", |
| 3553 | + "model_context_window_exceeded" |
3555 | 3554 | ] |
3556 | 3555 | }, |
3557 | 3556 | "String":{"type":"string"}, |
|
0 commit comments