@@ -101,7 +101,9 @@ def parse(
101
101
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
102
102
prediction : Optional [ChatCompletionPredictionContentParam ] | NotGiven = NOT_GIVEN ,
103
103
presence_penalty : Optional [float ] | NotGiven = NOT_GIVEN ,
104
+ prompt_cache_key : str | NotGiven = NOT_GIVEN ,
104
105
reasoning_effort : Optional [ReasoningEffort ] | NotGiven = NOT_GIVEN ,
106
+ safety_identifier : str | NotGiven = NOT_GIVEN ,
105
107
seed : Optional [int ] | NotGiven = NOT_GIVEN ,
106
108
service_tier : Optional [Literal ["auto" , "default" , "flex" , "scale" , "priority" ]] | NotGiven = NOT_GIVEN ,
107
109
stop : Union [Optional [str ], List [str ], None ] | NotGiven = NOT_GIVEN ,
@@ -197,8 +199,10 @@ def parser(raw_completion: ChatCompletion) -> ParsedChatCompletion[ResponseForma
197
199
"parallel_tool_calls" : parallel_tool_calls ,
198
200
"prediction" : prediction ,
199
201
"presence_penalty" : presence_penalty ,
202
+ "prompt_cache_key" : prompt_cache_key ,
200
203
"reasoning_effort" : reasoning_effort ,
201
204
"response_format" : _type_to_response_format (response_format ),
205
+ "safety_identifier" : safety_identifier ,
202
206
"seed" : seed ,
203
207
"service_tier" : service_tier ,
204
208
"stop" : stop ,
@@ -1378,7 +1382,9 @@ def stream(
1378
1382
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1379
1383
prediction : Optional [ChatCompletionPredictionContentParam ] | NotGiven = NOT_GIVEN ,
1380
1384
presence_penalty : Optional [float ] | NotGiven = NOT_GIVEN ,
1385
+ prompt_cache_key : str | NotGiven = NOT_GIVEN ,
1381
1386
reasoning_effort : Optional [ReasoningEffort ] | NotGiven = NOT_GIVEN ,
1387
+ safety_identifier : str | NotGiven = NOT_GIVEN ,
1382
1388
seed : Optional [int ] | NotGiven = NOT_GIVEN ,
1383
1389
service_tier : Optional [Literal ["auto" , "default" , "flex" , "scale" , "priority" ]] | NotGiven = NOT_GIVEN ,
1384
1390
stop : Union [Optional [str ], List [str ], None ] | NotGiven = NOT_GIVEN ,
@@ -1445,7 +1451,9 @@ def stream(
1445
1451
parallel_tool_calls = parallel_tool_calls ,
1446
1452
prediction = prediction ,
1447
1453
presence_penalty = presence_penalty ,
1454
+ prompt_cache_key = prompt_cache_key ,
1448
1455
reasoning_effort = reasoning_effort ,
1456
+ safety_identifier = safety_identifier ,
1449
1457
seed = seed ,
1450
1458
service_tier = service_tier ,
1451
1459
store = store ,
@@ -1514,7 +1522,9 @@ async def parse(
1514
1522
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
1515
1523
prediction : Optional [ChatCompletionPredictionContentParam ] | NotGiven = NOT_GIVEN ,
1516
1524
presence_penalty : Optional [float ] | NotGiven = NOT_GIVEN ,
1525
+ prompt_cache_key : str | NotGiven = NOT_GIVEN ,
1517
1526
reasoning_effort : Optional [ReasoningEffort ] | NotGiven = NOT_GIVEN ,
1527
+ safety_identifier : str | NotGiven = NOT_GIVEN ,
1518
1528
seed : Optional [int ] | NotGiven = NOT_GIVEN ,
1519
1529
service_tier : Optional [Literal ["auto" , "default" , "flex" , "scale" , "priority" ]] | NotGiven = NOT_GIVEN ,
1520
1530
stop : Union [Optional [str ], List [str ], None ] | NotGiven = NOT_GIVEN ,
@@ -1610,8 +1620,10 @@ def parser(raw_completion: ChatCompletion) -> ParsedChatCompletion[ResponseForma
1610
1620
"parallel_tool_calls" : parallel_tool_calls ,
1611
1621
"prediction" : prediction ,
1612
1622
"presence_penalty" : presence_penalty ,
1623
+ "prompt_cache_key" : prompt_cache_key ,
1613
1624
"reasoning_effort" : reasoning_effort ,
1614
1625
"response_format" : _type_to_response_format (response_format ),
1626
+ "safety_identifier" : safety_identifier ,
1615
1627
"seed" : seed ,
1616
1628
"service_tier" : service_tier ,
1617
1629
"store" : store ,
@@ -2791,7 +2803,9 @@ def stream(
2791
2803
parallel_tool_calls : bool | NotGiven = NOT_GIVEN ,
2792
2804
prediction : Optional [ChatCompletionPredictionContentParam ] | NotGiven = NOT_GIVEN ,
2793
2805
presence_penalty : Optional [float ] | NotGiven = NOT_GIVEN ,
2806
+ prompt_cache_key : str | NotGiven = NOT_GIVEN ,
2794
2807
reasoning_effort : Optional [ReasoningEffort ] | NotGiven = NOT_GIVEN ,
2808
+ safety_identifier : str | NotGiven = NOT_GIVEN ,
2795
2809
seed : Optional [int ] | NotGiven = NOT_GIVEN ,
2796
2810
service_tier : Optional [Literal ["auto" , "default" , "flex" , "scale" , "priority" ]] | NotGiven = NOT_GIVEN ,
2797
2811
stop : Union [Optional [str ], List [str ], None ] | NotGiven = NOT_GIVEN ,
@@ -2859,7 +2873,9 @@ def stream(
2859
2873
parallel_tool_calls = parallel_tool_calls ,
2860
2874
prediction = prediction ,
2861
2875
presence_penalty = presence_penalty ,
2876
+ prompt_cache_key = prompt_cache_key ,
2862
2877
reasoning_effort = reasoning_effort ,
2878
+ safety_identifier = safety_identifier ,
2863
2879
seed = seed ,
2864
2880
service_tier = service_tier ,
2865
2881
stop = stop ,
0 commit comments