Commit b204d41e
Changed files (2)
src
openai
resources
chat
completions
responses
src/openai/resources/chat/completions/completions.py
@@ -101,7 +101,9 @@ class Completions(SyncAPIResource):
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning_effort: Optional[ReasoningEffort] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
stop: Union[Optional[str], List[str], None] | NotGiven = NOT_GIVEN,
@@ -197,8 +199,10 @@ class Completions(SyncAPIResource):
"parallel_tool_calls": parallel_tool_calls,
"prediction": prediction,
"presence_penalty": presence_penalty,
+ "prompt_cache_key": prompt_cache_key,
"reasoning_effort": reasoning_effort,
"response_format": _type_to_response_format(response_format),
+ "safety_identifier": safety_identifier,
"seed": seed,
"service_tier": service_tier,
"stop": stop,
@@ -1378,7 +1382,9 @@ class Completions(SyncAPIResource):
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning_effort: Optional[ReasoningEffort] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
stop: Union[Optional[str], List[str], None] | NotGiven = NOT_GIVEN,
@@ -1445,7 +1451,9 @@ class Completions(SyncAPIResource):
parallel_tool_calls=parallel_tool_calls,
prediction=prediction,
presence_penalty=presence_penalty,
+ prompt_cache_key=prompt_cache_key,
reasoning_effort=reasoning_effort,
+ safety_identifier=safety_identifier,
seed=seed,
service_tier=service_tier,
store=store,
@@ -1514,7 +1522,9 @@ class AsyncCompletions(AsyncAPIResource):
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning_effort: Optional[ReasoningEffort] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
stop: Union[Optional[str], List[str], None] | NotGiven = NOT_GIVEN,
@@ -1610,8 +1620,10 @@ class AsyncCompletions(AsyncAPIResource):
"parallel_tool_calls": parallel_tool_calls,
"prediction": prediction,
"presence_penalty": presence_penalty,
+ "prompt_cache_key": prompt_cache_key,
"reasoning_effort": reasoning_effort,
"response_format": _type_to_response_format(response_format),
+ "safety_identifier": safety_identifier,
"seed": seed,
"service_tier": service_tier,
"store": store,
@@ -2791,7 +2803,9 @@ class AsyncCompletions(AsyncAPIResource):
parallel_tool_calls: bool | NotGiven = NOT_GIVEN,
prediction: Optional[ChatCompletionPredictionContentParam] | NotGiven = NOT_GIVEN,
presence_penalty: Optional[float] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning_effort: Optional[ReasoningEffort] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
seed: Optional[int] | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
stop: Union[Optional[str], List[str], None] | NotGiven = NOT_GIVEN,
@@ -2859,7 +2873,9 @@ class AsyncCompletions(AsyncAPIResource):
parallel_tool_calls=parallel_tool_calls,
prediction=prediction,
presence_penalty=presence_penalty,
+ prompt_cache_key=prompt_cache_key,
reasoning_effort=reasoning_effort,
+ safety_identifier=safety_identifier,
seed=seed,
service_tier=service_tier,
stop=stop,
src/openai/resources/responses/responses.py
@@ -1001,7 +1001,9 @@ class Responses(SyncAPIResource):
parallel_tool_calls: Optional[bool] | NotGiven = NOT_GIVEN,
previous_response_id: Optional[str] | NotGiven = NOT_GIVEN,
prompt: Optional[ResponsePromptParam] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning: Optional[Reasoning] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
store: Optional[bool] | NotGiven = NOT_GIVEN,
stream: Optional[Literal[False]] | Literal[True] | NotGiven = NOT_GIVEN,
@@ -1053,7 +1055,9 @@ class Responses(SyncAPIResource):
"parallel_tool_calls": parallel_tool_calls,
"previous_response_id": previous_response_id,
"prompt": prompt,
+ "prompt_cache_key": prompt_cache_key,
"reasoning": reasoning,
+ "safety_identifier": safety_identifier,
"service_tier": service_tier,
"store": store,
"stream": stream,
@@ -2316,7 +2320,9 @@ class AsyncResponses(AsyncAPIResource):
parallel_tool_calls: Optional[bool] | NotGiven = NOT_GIVEN,
previous_response_id: Optional[str] | NotGiven = NOT_GIVEN,
prompt: Optional[ResponsePromptParam] | NotGiven = NOT_GIVEN,
+ prompt_cache_key: str | NotGiven = NOT_GIVEN,
reasoning: Optional[Reasoning] | NotGiven = NOT_GIVEN,
+ safety_identifier: str | NotGiven = NOT_GIVEN,
service_tier: Optional[Literal["auto", "default", "flex", "scale", "priority"]] | NotGiven = NOT_GIVEN,
store: Optional[bool] | NotGiven = NOT_GIVEN,
stream: Optional[Literal[False]] | Literal[True] | NotGiven = NOT_GIVEN,
@@ -2368,7 +2374,9 @@ class AsyncResponses(AsyncAPIResource):
"parallel_tool_calls": parallel_tool_calls,
"previous_response_id": previous_response_id,
"prompt": prompt,
+ "prompt_cache_key": prompt_cache_key,
"reasoning": reasoning,
+ "safety_identifier": safety_identifier,
"service_tier": service_tier,
"store": store,
"stream": stream,