Commit d4c2c7a3
Changed files (14)
src
openai
src/openai/resources/beta/threads/runs/runs.py
@@ -164,8 +164,9 @@ class Runs(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -314,8 +315,9 @@ class Runs(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -460,8 +462,9 @@ class Runs(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -1097,7 +1100,7 @@ class Runs(SyncAPIResource):
if is_given(poll_interval_ms):
extra_headers["X-Stainless-Custom-Poll-Interval"] = str(poll_interval_ms)
- terminal_states = {"requires_action", "cancelled", "completed", "failed", "expired"}
+ terminal_states = {"requires_action", "cancelled", "completed", "failed", "expired", "incomplete"}
while True:
response = self.with_raw_response.retrieve(
thread_id=thread_id,
@@ -1718,8 +1721,9 @@ class AsyncRuns(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -1868,8 +1872,9 @@ class AsyncRuns(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -2014,8 +2019,9 @@ class AsyncRuns(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -2653,7 +2659,7 @@ class AsyncRuns(AsyncAPIResource):
if is_given(poll_interval_ms):
extra_headers["X-Stainless-Custom-Poll-Interval"] = str(poll_interval_ms)
- terminal_states = {"requires_action", "cancelled", "completed", "failed", "expired"}
+ terminal_states = {"requires_action", "cancelled", "completed", "failed", "expired", "incomplete"}
while True:
response = await self.with_raw_response.retrieve(
thread_id=thread_id,
src/openai/resources/beta/threads/threads.py
@@ -341,8 +341,9 @@ class Threads(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -490,8 +491,9 @@ class Threads(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -635,8 +637,9 @@ class Threads(SyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -1331,8 +1334,9 @@ class AsyncThreads(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -1480,8 +1484,9 @@ class AsyncThreads(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -1625,8 +1630,9 @@ class AsyncThreads(AsyncAPIResource):
assistant will be used.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/resources/beta/assistants.py
@@ -110,8 +110,9 @@ class Assistants(SyncAPIResource):
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -254,8 +255,9 @@ class Assistants(SyncAPIResource):
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -497,8 +499,9 @@ class AsyncAssistants(AsyncAPIResource):
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -641,8 +644,9 @@ class AsyncAssistants(AsyncAPIResource):
name: The name of the assistant. The maximum length is 256 characters.
response_format: Specifies the format that the model must output. Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/resources/batches.py
@@ -40,7 +40,7 @@ class Batches(SyncAPIResource):
self,
*,
completion_window: Literal["24h"],
- endpoint: Literal["/v1/chat/completions", "/v1/embeddings"],
+ endpoint: Literal["/v1/chat/completions", "/v1/embeddings", "/v1/completions"],
input_file_id: str,
metadata: Optional[Dict[str, str]] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -58,7 +58,9 @@ class Batches(SyncAPIResource):
is supported.
endpoint: The endpoint to be used for all requests in the batch. Currently
- `/v1/chat/completions` and `/v1/embeddings` are supported.
+ `/v1/chat/completions`, `/v1/embeddings`, and `/v1/completions` are supported.
+ Note that `/v1/embeddings` batches are also restricted to a maximum of 50,000
+ embedding inputs across all requests in the batch.
input_file_id: The ID of an uploaded file that contains requests for the new batch.
@@ -67,7 +69,8 @@ class Batches(SyncAPIResource):
Your input file must be formatted as a
[JSONL file](https://platform.openai.com/docs/api-reference/batch/requestInput),
- and must be uploaded with the purpose `batch`.
+ and must be uploaded with the purpose `batch`. The file can contain up to 50,000
+ requests, and can be up to 100 MB in size.
metadata: Optional custom metadata for the batch.
@@ -228,7 +231,7 @@ class AsyncBatches(AsyncAPIResource):
self,
*,
completion_window: Literal["24h"],
- endpoint: Literal["/v1/chat/completions", "/v1/embeddings"],
+ endpoint: Literal["/v1/chat/completions", "/v1/embeddings", "/v1/completions"],
input_file_id: str,
metadata: Optional[Dict[str, str]] | NotGiven = NOT_GIVEN,
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
@@ -246,7 +249,9 @@ class AsyncBatches(AsyncAPIResource):
is supported.
endpoint: The endpoint to be used for all requests in the batch. Currently
- `/v1/chat/completions` and `/v1/embeddings` are supported.
+ `/v1/chat/completions`, `/v1/embeddings`, and `/v1/completions` are supported.
+ Note that `/v1/embeddings` batches are also restricted to a maximum of 50,000
+ embedding inputs across all requests in the batch.
input_file_id: The ID of an uploaded file that contains requests for the new batch.
@@ -255,7 +260,8 @@ class AsyncBatches(AsyncAPIResource):
Your input file must be formatted as a
[JSONL file](https://platform.openai.com/docs/api-reference/batch/requestInput),
- and must be uploaded with the purpose `batch`.
+ and must be uploaded with the purpose `batch`. The file can contain up to 50,000
+ requests, and can be up to 100 MB in size.
metadata: Optional custom metadata for the batch.
src/openai/resources/files.py
@@ -62,14 +62,18 @@ class Files(SyncAPIResource):
) -> FileObject:
"""Upload a file that can be used across various endpoints.
- The size of all the
- files uploaded by one organization can be up to 100 GB.
+ Individual files can be
+ up to 512 MB, and the size of all files uploaded by one organization can be up
+ to 100 GB.
- The size of individual files can be a maximum of 512 MB or 2 million tokens for
- Assistants. See the
- [Assistants Tools guide](https://platform.openai.com/docs/assistants/tools) to
- learn more about the types of files supported. The Fine-tuning API only supports
- `.jsonl` files.
+ The Assistants API supports files up to 2 million tokens and of specific file
+ types. See the
+ [Assistants Tools guide](https://platform.openai.com/docs/assistants/tools) for
+ details.
+
+ The Fine-tuning API only supports `.jsonl` files.
+
+ The Batch API only supports `.jsonl` files up to 100 MB in size.
Please [contact us](https://help.openai.com/) if you need to increase these
storage limits.
@@ -335,14 +339,18 @@ class AsyncFiles(AsyncAPIResource):
) -> FileObject:
"""Upload a file that can be used across various endpoints.
- The size of all the
- files uploaded by one organization can be up to 100 GB.
+ Individual files can be
+ up to 512 MB, and the size of all files uploaded by one organization can be up
+ to 100 GB.
+
+ The Assistants API supports files up to 2 million tokens and of specific file
+ types. See the
+ [Assistants Tools guide](https://platform.openai.com/docs/assistants/tools) for
+ details.
+
+ The Fine-tuning API only supports `.jsonl` files.
- The size of individual files can be a maximum of 512 MB or 2 million tokens for
- Assistants. See the
- [Assistants Tools guide](https://platform.openai.com/docs/assistants/tools) to
- learn more about the types of files supported. The Fine-tuning API only supports
- `.jsonl` files.
+ The Batch API only supports `.jsonl` files up to 100 MB in size.
Please [contact us](https://help.openai.com/) if you need to increase these
storage limits.
src/openai/types/beta/threads/run.py
@@ -160,9 +160,9 @@ class Run(BaseModel):
response_format: Optional[AssistantResponseFormatOption] = None
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
@@ -182,8 +182,8 @@ class Run(BaseModel):
status: RunStatus
"""
The status of the run, which can be either `queued`, `in_progress`,
- `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`, or
- `expired`.
+ `requires_action`, `cancelling`, `cancelled`, `failed`, `completed`,
+ `incomplete`, or `expired`.
"""
thread_id: str
src/openai/types/beta/threads/run_create_params.py
@@ -110,9 +110,9 @@ class RunCreateParamsBase(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/types/beta/threads/run_status.py
@@ -5,5 +5,13 @@ from typing_extensions import Literal
__all__ = ["RunStatus"]
RunStatus = Literal[
- "queued", "in_progress", "requires_action", "cancelling", "cancelled", "failed", "completed", "expired"
+ "queued",
+ "in_progress",
+ "requires_action",
+ "cancelling",
+ "cancelled",
+ "failed",
+ "completed",
+ "incomplete",
+ "expired",
]
src/openai/types/beta/assistant.py
@@ -85,9 +85,9 @@ class Assistant(BaseModel):
response_format: Optional[AssistantResponseFormatOption] = None
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/types/beta/assistant_create_params.py
@@ -77,9 +77,9 @@ class AssistantCreateParams(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/types/beta/assistant_update_params.py
@@ -45,9 +45,9 @@ class AssistantUpdateParams(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/types/beta/thread_create_and_run_params.py
@@ -108,9 +108,9 @@ class ThreadCreateAndRunParamsBase(TypedDict, total=False):
response_format: Optional[AssistantResponseFormatOptionParam]
"""Specifies the format that the model must output.
- Compatible with
- [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo) and
- all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
+ Compatible with [GPT-4o](https://platform.openai.com/docs/models/gpt-4o),
+ [GPT-4 Turbo](https://platform.openai.com/docs/models/gpt-4-turbo-and-gpt-4),
+ and all GPT-3.5 Turbo models since `gpt-3.5-turbo-1106`.
Setting to `{ "type": "json_object" }` enables JSON mode, which guarantees the
message the model generates is valid JSON.
src/openai/types/batch_create_params.py
@@ -15,10 +15,12 @@ class BatchCreateParams(TypedDict, total=False):
Currently only `24h` is supported.
"""
- endpoint: Required[Literal["/v1/chat/completions", "/v1/embeddings"]]
+ endpoint: Required[Literal["/v1/chat/completions", "/v1/embeddings", "/v1/completions"]]
"""The endpoint to be used for all requests in the batch.
- Currently `/v1/chat/completions` and `/v1/embeddings` are supported.
+ Currently `/v1/chat/completions`, `/v1/embeddings`, and `/v1/completions` are
+ supported. Note that `/v1/embeddings` batches are also restricted to a maximum
+ of 50,000 embedding inputs across all requests in the batch.
"""
input_file_id: Required[str]
@@ -29,7 +31,8 @@ class BatchCreateParams(TypedDict, total=False):
Your input file must be formatted as a
[JSONL file](https://platform.openai.com/docs/api-reference/batch/requestInput),
- and must be uploaded with the purpose `batch`.
+ and must be uploaded with the purpose `batch`. The file can contain up to 50,000
+ requests, and can be up to 100 MB in size.
"""
metadata: Optional[Dict[str, str]]
.stats.yml
@@ -1,2 +1,2 @@
configured_endpoints: 64
-openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-47007cc1aa5bc7b74107a99b377925978a0bd376ed67bdae724e80d5d0b63d57.yml
+openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/openai-363dd904e5d6e65b3a323fc88e6b502fb23a6aa319be219273e3ee47c7530993.yml