Commit 2bfec1a2
Changed files (22)
src
openai
resources
beta
chat
fine_tuning
jobs
tests
api_resources
src/openai/resources/beta/chat/completions.py
@@ -42,6 +42,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -121,6 +122,7 @@ class Completions(SyncAPIResource):
functions=functions,
logit_bias=logit_bias,
logprobs=logprobs,
+ max_completion_tokens=max_completion_tokens,
max_tokens=max_tokens,
n=n,
parallel_tool_calls=parallel_tool_calls,
@@ -157,6 +159,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -216,6 +219,7 @@ class Completions(SyncAPIResource):
functions=functions,
logit_bias=logit_bias,
logprobs=logprobs,
+ max_completion_tokens=max_completion_tokens,
max_tokens=max_tokens,
n=n,
parallel_tool_calls=parallel_tool_calls,
@@ -254,6 +258,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -333,6 +338,7 @@ class AsyncCompletions(AsyncAPIResource):
functions=functions,
logit_bias=logit_bias,
logprobs=logprobs,
+ max_completion_tokens=max_completion_tokens,
max_tokens=max_tokens,
n=n,
parallel_tool_calls=parallel_tool_calls,
@@ -369,6 +375,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -429,6 +436,7 @@ class AsyncCompletions(AsyncAPIResource):
functions=functions,
logit_bias=logit_bias,
logprobs=logprobs,
+ max_completion_tokens=max_completion_tokens,
max_tokens=max_tokens,
n=n,
parallel_tool_calls=parallel_tool_calls,
src/openai/resources/beta/threads/runs/runs.py
@@ -167,11 +167,11 @@ class Runs(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -311,11 +311,11 @@ class Runs(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -451,11 +451,11 @@ class Runs(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1529,11 +1529,11 @@ class AsyncRuns(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1673,11 +1673,11 @@ class AsyncRuns(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1813,11 +1813,11 @@ class AsyncRuns(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/resources/beta/threads/threads.py
@@ -335,11 +335,11 @@ class Threads(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -469,11 +469,11 @@ class Threads(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -599,11 +599,11 @@ class Threads(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1169,11 +1169,11 @@ class AsyncThreads(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1303,11 +1303,11 @@ class AsyncThreads(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1433,11 +1433,11 @@ class AsyncThreads(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/resources/beta/assistants.py
@@ -100,11 +100,11 @@ class Assistants(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -250,11 +250,11 @@ class Assistants(SyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -486,11 +486,11 @@ class AsyncAssistants(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -636,11 +636,11 @@ class AsyncAssistants(AsyncAPIResource):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/resources/chat/completions.py
@@ -64,6 +64,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -132,13 +133,17 @@ class Completions(SyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -161,11 +166,11 @@ class Completions(SyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -185,8 +190,11 @@ class Completions(SyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -261,6 +269,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -335,13 +344,17 @@ class Completions(SyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -364,11 +377,11 @@ class Completions(SyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -388,8 +401,11 @@ class Completions(SyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -457,6 +473,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -531,13 +548,17 @@ class Completions(SyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -560,11 +581,11 @@ class Completions(SyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -584,8 +605,11 @@ class Completions(SyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -652,6 +676,7 @@ class Completions(SyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -687,6 +712,7 @@ class Completions(SyncAPIResource):
"functions": functions,
"logit_bias": logit_bias,
"logprobs": logprobs,
+ "max_completion_tokens": max_completion_tokens,
"max_tokens": max_tokens,
"n": n,
"parallel_tool_calls": parallel_tool_calls,
@@ -746,6 +772,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -814,13 +841,17 @@ class AsyncCompletions(AsyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -843,11 +874,11 @@ class AsyncCompletions(AsyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -867,8 +898,11 @@ class AsyncCompletions(AsyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -943,6 +977,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -1017,13 +1052,17 @@ class AsyncCompletions(AsyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -1046,11 +1085,11 @@ class AsyncCompletions(AsyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1070,8 +1109,11 @@ class AsyncCompletions(AsyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -1139,6 +1181,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -1213,13 +1256,17 @@ class AsyncCompletions(AsyncAPIResource):
returns the log probabilities of each output token returned in the `content` of
`message`.
+ max_completion_tokens: An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+
max_tokens: The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
n: How many chat completion choices to generate for each input message. Note that
you will be charged based on the number of generated tokens across all of the
@@ -1242,11 +1289,11 @@ class AsyncCompletions(AsyncAPIResource):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -1266,8 +1313,11 @@ class AsyncCompletions(AsyncAPIResource):
service_tier: Specifies the latency tier to use for processing the request. This parameter is
relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
@@ -1334,6 +1384,7 @@ class AsyncCompletions(AsyncAPIResource):
functions: Iterable[completion_create_params.Function] | NotGiven = NOT_GIVEN,
logit_bias: Optional[Dict[str, int]] | NotGiven = NOT_GIVEN,
logprobs: Optional[bool] | NotGiven = NOT_GIVEN,
+ max_completion_tokens: Optional[int] | NotGiven = NOT_GIVEN,
max_tokens: Optional[int] | NotGiven = NOT_GIVEN,
n: Optional[int] | NotGiven = NOT_GIVEN,
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
@@ -1369,6 +1420,7 @@ class AsyncCompletions(AsyncAPIResource):
"functions": functions,
"logit_bias": logit_bias,
"logprobs": logprobs,
+ "max_completion_tokens": max_completion_tokens,
"max_tokens": max_tokens,
"n": n,
"parallel_tool_calls": parallel_tool_calls,
src/openai/resources/fine_tuning/jobs/jobs.py
@@ -114,7 +114,7 @@ class Jobs(SyncAPIResource):
job parameters should produce the same results, but may differ in rare cases. If
a seed is not specified, one will be generated for you.
- suffix: A string of up to 18 characters that will be added to your fine-tuned model
+ suffix: A string of up to 64 characters that will be added to your fine-tuned model
name.
For example, a `suffix` of "custom-model-name" would produce a model name like
@@ -405,7 +405,7 @@ class AsyncJobs(AsyncAPIResource):
job parameters should produce the same results, but may differ in rare cases. If
a seed is not specified, one will be generated for you.
- suffix: A string of up to 18 characters that will be added to your fine-tuned model
+ suffix: A string of up to 64 characters that will be added to your fine-tuned model
name.
For example, a `suffix` of "custom-model-name" would produce a model name like
src/openai/types/beta/threads/run.py
@@ -172,11 +172,11 @@ class Run(BaseModel):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/beta/threads/run_create_params.py
@@ -111,11 +111,11 @@ class RunCreateParamsBase(TypedDict, total=False):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/beta/assistant.py
@@ -90,11 +90,11 @@ class Assistant(BaseModel):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/beta/assistant_create_params.py
@@ -58,11 +58,11 @@ class AssistantCreateParams(TypedDict, total=False):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/beta/assistant_update_params.py
@@ -50,11 +50,11 @@ class AssistantUpdateParams(TypedDict, total=False):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/beta/file_search_tool.py
@@ -9,16 +9,16 @@ __all__ = ["FileSearchTool", "FileSearch", "FileSearchRankingOptions"]
class FileSearchRankingOptions(BaseModel):
- ranker: Optional[Literal["auto", "default_2024_08_21"]] = None
- """The ranker to use for the file search.
+ score_threshold: float
+ """The score threshold for the file search.
- If not specified will use the `auto` ranker.
+ All values must be a floating point number between 0 and 1.
"""
- score_threshold: Optional[float] = None
- """The score threshold for the file search.
+ ranker: Optional[Literal["auto", "default_2024_08_21"]] = None
+ """The ranker to use for the file search.
- All values must be a floating point number between 0 and 1.
+ If not specified will use the `auto` ranker.
"""
@@ -38,6 +38,9 @@ class FileSearch(BaseModel):
ranking_options: Optional[FileSearchRankingOptions] = None
"""The ranking options for the file search.
+ If not specified, the file search tool will use the `auto` ranker and a
+ score_threshold of 0.
+
See the
[file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
for more information.
src/openai/types/beta/file_search_tool_param.py
@@ -8,16 +8,16 @@ __all__ = ["FileSearchToolParam", "FileSearch", "FileSearchRankingOptions"]
class FileSearchRankingOptions(TypedDict, total=False):
- ranker: Literal["auto", "default_2024_08_21"]
- """The ranker to use for the file search.
+ score_threshold: Required[float]
+ """The score threshold for the file search.
- If not specified will use the `auto` ranker.
+ All values must be a floating point number between 0 and 1.
"""
- score_threshold: float
- """The score threshold for the file search.
+ ranker: Literal["auto", "default_2024_08_21"]
+ """The ranker to use for the file search.
- All values must be a floating point number between 0 and 1.
+ If not specified will use the `auto` ranker.
"""
@@ -37,6 +37,9 @@ class FileSearch(TypedDict, total=False):
ranking_options: FileSearchRankingOptions
"""The ranking options for the file search.
+ If not specified, the file search tool will use the `auto` ranker and a
+ score_threshold of 0.
+
See the
[file search tool documentation](https://platform.openai.com/docs/assistants/tools/file-search/customizing-file-search-settings)
for more information.
src/openai/types/beta/thread_create_and_run_params.py
@@ -98,11 +98,11 @@ class ThreadCreateAndRunParamsBase(TypedDict, total=False):
and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
src/openai/types/chat/completion_create_params.py
@@ -87,15 +87,22 @@ class CompletionCreateParamsBase(TypedDict, total=False):
`content` of `message`.
"""
+ max_completion_tokens: Optional[int]
+ """
+ An upper bound for the number of tokens that can be generated for a completion,
+ including visible output tokens and
+ [reasoning tokens](https://platform.openai.com/docs/guides/reasoning).
+ """
+
max_tokens: Optional[int]
"""
The maximum number of [tokens](/tokenizer) that can be generated in the chat
- completion.
+ completion. This value can be used to control
+ [costs](https://openai.com/api/pricing/) for text generated via API.
- The total length of input tokens and generated tokens is limited by the model's
- context length.
- [Example Python code](https://cookbook.openai.com/examples/how_to_count_tokens_with_tiktoken)
- for counting tokens.
+ This value is now deprecated in favor of `max_completion_tokens`, and is not
+ compatible with
+ [o1 series models](https://platform.openai.com/docs/guides/reasoning).
"""
n: Optional[int]
@@ -130,11 +137,11 @@ class CompletionCreateParamsBase(TypedDict, total=False):
all GPT-3.5 Turbo models newer than `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_schema", "json_schema": {...} }` enables Structured
- Outputs which guarantees the model will match your supplied JSON schema. Learn
- more in the
+ Outputs which ensures the model will match your supplied JSON schema. Learn more
+ in the
[Structured Outputs guide](https://platform.openai.com/docs/guides/structured-outputs).
- Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
+ Setting to `{ "type": "json_object" }` enables JSON mode, which ensures the
message the model generates is valid JSON.
**Important:** when using JSON mode, you **must** also instruct the model to
@@ -160,8 +167,11 @@ class CompletionCreateParamsBase(TypedDict, total=False):
This parameter is relevant for customers subscribed to the scale tier service:
- - If set to 'auto', the system will utilize scale tier credits until they are
- exhausted.
+ - If set to 'auto', and the Project is Scale tier enabled, the system will
+ utilize scale tier credits until they are exhausted.
+ - If set to 'auto', and the Project is not Scale tier enabled, the request will
+ be processed using the default service tier with a lower uptime SLA and no
+ latency guarentee.
- If set to 'default', the request will be processed using the default service
tier with a lower uptime SLA and no latency guarentee.
- When not set, the default behavior is 'auto'.
src/openai/types/fine_tuning/job_create_params.py
@@ -50,7 +50,7 @@ class JobCreateParams(TypedDict, total=False):
suffix: Optional[str]
"""
- A string of up to 18 characters that will be added to your fine-tuned model
+ A string of up to 64 characters that will be added to your fine-tuned model
name.
For example, a `suffix` of "custom-model-name" would produce a model name like
src/openai/types/chat_model.py
@@ -5,9 +5,14 @@ from typing_extensions import Literal, TypeAlias
__all__ = ["ChatModel"]
ChatModel: TypeAlias = Literal[
+ "o1-preview",
+ "o1-preview-2024-09-12",
+ "o1-mini",
+ "o1-mini-2024-09-12",
"gpt-4o",
- "gpt-4o-2024-05-13",
"gpt-4o-2024-08-06",
+ "gpt-4o-2024-05-13",
+ "chatgpt-4o-latest",
"gpt-4o-mini",
"gpt-4o-mini-2024-07-18",
"gpt-4-turbo",
src/openai/types/completion_usage.py
@@ -1,9 +1,15 @@
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
+from typing import Optional
from .._models import BaseModel
-__all__ = ["CompletionUsage"]
+__all__ = ["CompletionUsage", "CompletionTokensDetails"]
+
+
+class CompletionTokensDetails(BaseModel):
+ reasoning_tokens: Optional[int] = None
+ """Tokens generated by the model for reasoning."""
class CompletionUsage(BaseModel):
@@ -15,3 +21,6 @@ class CompletionUsage(BaseModel):
total_tokens: int
"""Total number of tokens used in the request (prompt + completion)."""
+
+ completion_tokens_details: Optional[CompletionTokensDetails] = None
+ """Breakdown of tokens used in a completion."""
tests/api_resources/chat/test_completions.py
@@ -55,6 +55,7 @@ class TestCompletions:
],
logit_bias={"foo": 0},
logprobs=True,
+ max_completion_tokens=0,
max_tokens=0,
n=1,
parallel_tool_calls=True,
@@ -175,6 +176,7 @@ class TestCompletions:
],
logit_bias={"foo": 0},
logprobs=True,
+ max_completion_tokens=0,
max_tokens=0,
n=1,
parallel_tool_calls=True,
@@ -314,6 +316,7 @@ class TestAsyncCompletions:
],
logit_bias={"foo": 0},
logprobs=True,
+ max_completion_tokens=0,
max_tokens=0,
n=1,
parallel_tool_calls=True,
@@ -434,6 +437,7 @@ class TestAsyncCompletions:
],
logit_bias={"foo": 0},
logprobs=True,
+ max_completion_tokens=0,
max_tokens=0,
n=1,
parallel_tool_calls=True,
tests/lib/chat/test_completions.py
@@ -74,7 +74,7 @@ Francisco, please check a reliable weather website or app.",
object='chat.completion',
service_tier=None,
system_fingerprint='fp_845eaabc1f',
- usage=CompletionUsage(completion_tokens=28, prompt_tokens=14, total_tokens=42)
+ usage=CompletionUsage(completion_tokens=28, completion_tokens_details=None, prompt_tokens=14, total_tokens=42)
)
"""
)
@@ -129,7 +129,7 @@ ParsedChatCompletion[Location](
object='chat.completion',
service_tier=None,
system_fingerprint='fp_2a322c9ffc',
- usage=CompletionUsage(completion_tokens=14, prompt_tokens=17, total_tokens=31)
+ usage=CompletionUsage(completion_tokens=14, completion_tokens_details=None, prompt_tokens=17, total_tokens=31)
)
"""
)
@@ -186,7 +186,7 @@ ParsedChatCompletion[Location](
object='chat.completion',
service_tier=None,
system_fingerprint='fp_2a322c9ffc',
- usage=CompletionUsage(completion_tokens=14, prompt_tokens=17, total_tokens=31)
+ usage=CompletionUsage(completion_tokens=14, completion_tokens_details=None, prompt_tokens=17, total_tokens=31)
)
"""
)
@@ -368,7 +368,7 @@ ParsedChatCompletion[CalendarEvent](
object='chat.completion',
service_tier=None,
system_fingerprint='fp_2a322c9ffc',
- usage=CompletionUsage(completion_tokens=17, prompt_tokens=32, total_tokens=49)
+ usage=CompletionUsage(completion_tokens=17, completion_tokens_details=None, prompt_tokens=32, total_tokens=49)
)
"""
)
tests/lib/chat/test_completions_streaming.py
@@ -155,7 +155,7 @@ ParsedChatCompletion[Location](
object='chat.completion',
service_tier=None,
system_fingerprint='fp_845eaabc1f',
- usage=CompletionUsage(completion_tokens=14, prompt_tokens=17, total_tokens=31)
+ usage=CompletionUsage(completion_tokens=14, completion_tokens_details=None, prompt_tokens=17, total_tokens=31)
)
"""
)
.stats.yml
@@ -1,2 +1,2 @@
configured_endpoints: 68
-openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-85a85e0c08de456441431c0ae4e9c078cc8f9748c29430b9a9058340db6389ee.yml
+openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-501122aa32adaa2abb3d4487880ab9cdf2141addce2e6c3d1bd9bb6b44c318a8.yml