Skip to content

Commit f8a6434

Browse files
Merge pull request #14942 from danielmklein/main
Fix inconsistent token configs for gpt-5 models
2 parents 0350a44 + de795a4 commit f8a6434

File tree

2 files changed

+26
-26
lines changed

2 files changed

+26
-26
lines changed

litellm/model_prices_and_context_window_backup.json

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -12735,7 +12735,7 @@
1273512735
"input_cost_per_token_flex": 6.25e-07,
1273612736
"input_cost_per_token_priority": 2.5e-06,
1273712737
"litellm_provider": "openai",
12738-
"max_input_tokens": 400000,
12738+
"max_input_tokens": 272000,
1273912739
"max_output_tokens": 128000,
1274012740
"max_tokens": 128000,
1274112741
"mode": "chat",
@@ -12773,7 +12773,7 @@
1277312773
"input_cost_per_token_flex": 6.25e-07,
1277412774
"input_cost_per_token_priority": 2.5e-06,
1277512775
"litellm_provider": "openai",
12776-
"max_input_tokens": 400000,
12776+
"max_input_tokens": 272000,
1277712777
"max_output_tokens": 128000,
1277812778
"max_tokens": 128000,
1277912779
"mode": "chat",
@@ -12808,7 +12808,7 @@
1280812808
"cache_read_input_token_cost": 1.25e-07,
1280912809
"input_cost_per_token": 1.25e-06,
1281012810
"litellm_provider": "openai",
12811-
"max_input_tokens": 400000,
12811+
"max_input_tokens": 272000,
1281212812
"max_output_tokens": 128000,
1281312813
"max_tokens": 128000,
1281412814
"mode": "chat",
@@ -12840,7 +12840,7 @@
1284012840
"cache_read_input_token_cost": 1.25e-07,
1284112841
"input_cost_per_token": 1.25e-06,
1284212842
"litellm_provider": "openai",
12843-
"max_input_tokens": 400000,
12843+
"max_input_tokens": 272000,
1284412844
"max_output_tokens": 128000,
1284512845
"max_tokens": 128000,
1284612846
"mode": "chat",
@@ -12906,7 +12906,7 @@
1290612906
"input_cost_per_token_flex": 1.25e-07,
1290712907
"input_cost_per_token_priority": 4.5e-07,
1290812908
"litellm_provider": "openai",
12909-
"max_input_tokens": 400000,
12909+
"max_input_tokens": 272000,
1291012910
"max_output_tokens": 128000,
1291112911
"max_tokens": 128000,
1291212912
"mode": "chat",
@@ -12944,7 +12944,7 @@
1294412944
"input_cost_per_token_flex": 1.25e-07,
1294512945
"input_cost_per_token_priority": 4.5e-07,
1294612946
"litellm_provider": "openai",
12947-
"max_input_tokens": 400000,
12947+
"max_input_tokens": 272000,
1294812948
"max_output_tokens": 128000,
1294912949
"max_tokens": 128000,
1295012950
"mode": "chat",
@@ -12981,7 +12981,7 @@
1298112981
"input_cost_per_token_flex": 2.5e-08,
1298212982
"input_cost_per_token_priority": 2.5e-06,
1298312983
"litellm_provider": "openai",
12984-
"max_input_tokens": 400000,
12984+
"max_input_tokens": 272000,
1298512985
"max_output_tokens": 128000,
1298612986
"max_tokens": 128000,
1298712987
"mode": "chat",
@@ -13016,7 +13016,7 @@
1301613016
"input_cost_per_token": 5e-08,
1301713017
"input_cost_per_token_flex": 2.5e-08,
1301813018
"litellm_provider": "openai",
13019-
"max_input_tokens": 400000,
13019+
"max_input_tokens": 272000,
1302013020
"max_output_tokens": 128000,
1302113021
"max_tokens": 128000,
1302213022
"mode": "chat",
@@ -17210,7 +17210,7 @@
1721017210
"cache_read_input_token_cost": 1.25e-07,
1721117211
"input_cost_per_token": 1.25e-06,
1721217212
"litellm_provider": "openrouter",
17213-
"max_input_tokens": 400000,
17213+
"max_input_tokens": 272000,
1721417214
"max_output_tokens": 128000,
1721517215
"max_tokens": 128000,
1721617216
"mode": "chat",
@@ -17229,7 +17229,7 @@
1722917229
"cache_read_input_token_cost": 1.25e-07,
1723017230
"input_cost_per_token": 1.25e-06,
1723117231
"litellm_provider": "openrouter",
17232-
"max_input_tokens": 400000,
17232+
"max_input_tokens": 272000,
1723317233
"max_output_tokens": 128000,
1723417234
"max_tokens": 128000,
1723517235
"mode": "chat",
@@ -17248,7 +17248,7 @@
1724817248
"cache_read_input_token_cost": 1.25e-07,
1724917249
"input_cost_per_token": 1.25e-06,
1725017250
"litellm_provider": "openrouter",
17251-
"max_input_tokens": 400000,
17251+
"max_input_tokens": 272000,
1725217252
"max_output_tokens": 128000,
1725317253
"max_tokens": 128000,
1725417254
"mode": "chat",
@@ -17267,7 +17267,7 @@
1726717267
"cache_read_input_token_cost": 2.5e-08,
1726817268
"input_cost_per_token": 2.5e-07,
1726917269
"litellm_provider": "openrouter",
17270-
"max_input_tokens": 400000,
17270+
"max_input_tokens": 272000,
1727117271
"max_output_tokens": 128000,
1727217272
"max_tokens": 128000,
1727317273
"mode": "chat",
@@ -17286,7 +17286,7 @@
1728617286
"cache_read_input_token_cost": 5e-09,
1728717287
"input_cost_per_token": 5e-08,
1728817288
"litellm_provider": "openrouter",
17289-
"max_input_tokens": 400000,
17289+
"max_input_tokens": 272000,
1729017290
"max_output_tokens": 128000,
1729117291
"max_tokens": 128000,
1729217292
"mode": "chat",

model_prices_and_context_window.json

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -12735,7 +12735,7 @@
1273512735
"input_cost_per_token_flex": 6.25e-07,
1273612736
"input_cost_per_token_priority": 2.5e-06,
1273712737
"litellm_provider": "openai",
12738-
"max_input_tokens": 400000,
12738+
"max_input_tokens": 272000,
1273912739
"max_output_tokens": 128000,
1274012740
"max_tokens": 128000,
1274112741
"mode": "chat",
@@ -12773,7 +12773,7 @@
1277312773
"input_cost_per_token_flex": 6.25e-07,
1277412774
"input_cost_per_token_priority": 2.5e-06,
1277512775
"litellm_provider": "openai",
12776-
"max_input_tokens": 400000,
12776+
"max_input_tokens": 272000,
1277712777
"max_output_tokens": 128000,
1277812778
"max_tokens": 128000,
1277912779
"mode": "chat",
@@ -12808,7 +12808,7 @@
1280812808
"cache_read_input_token_cost": 1.25e-07,
1280912809
"input_cost_per_token": 1.25e-06,
1281012810
"litellm_provider": "openai",
12811-
"max_input_tokens": 400000,
12811+
"max_input_tokens": 272000,
1281212812
"max_output_tokens": 128000,
1281312813
"max_tokens": 128000,
1281412814
"mode": "chat",
@@ -12840,7 +12840,7 @@
1284012840
"cache_read_input_token_cost": 1.25e-07,
1284112841
"input_cost_per_token": 1.25e-06,
1284212842
"litellm_provider": "openai",
12843-
"max_input_tokens": 400000,
12843+
"max_input_tokens": 272000,
1284412844
"max_output_tokens": 128000,
1284512845
"max_tokens": 128000,
1284612846
"mode": "chat",
@@ -12906,7 +12906,7 @@
1290612906
"input_cost_per_token_flex": 1.25e-07,
1290712907
"input_cost_per_token_priority": 4.5e-07,
1290812908
"litellm_provider": "openai",
12909-
"max_input_tokens": 400000,
12909+
"max_input_tokens": 272000,
1291012910
"max_output_tokens": 128000,
1291112911
"max_tokens": 128000,
1291212912
"mode": "chat",
@@ -12944,7 +12944,7 @@
1294412944
"input_cost_per_token_flex": 1.25e-07,
1294512945
"input_cost_per_token_priority": 4.5e-07,
1294612946
"litellm_provider": "openai",
12947-
"max_input_tokens": 400000,
12947+
"max_input_tokens": 272000,
1294812948
"max_output_tokens": 128000,
1294912949
"max_tokens": 128000,
1295012950
"mode": "chat",
@@ -12981,7 +12981,7 @@
1298112981
"input_cost_per_token_flex": 2.5e-08,
1298212982
"input_cost_per_token_priority": 2.5e-06,
1298312983
"litellm_provider": "openai",
12984-
"max_input_tokens": 400000,
12984+
"max_input_tokens": 272000,
1298512985
"max_output_tokens": 128000,
1298612986
"max_tokens": 128000,
1298712987
"mode": "chat",
@@ -13016,7 +13016,7 @@
1301613016
"input_cost_per_token": 5e-08,
1301713017
"input_cost_per_token_flex": 2.5e-08,
1301813018
"litellm_provider": "openai",
13019-
"max_input_tokens": 400000,
13019+
"max_input_tokens": 272000,
1302013020
"max_output_tokens": 128000,
1302113021
"max_tokens": 128000,
1302213022
"mode": "chat",
@@ -17210,7 +17210,7 @@
1721017210
"cache_read_input_token_cost": 1.25e-07,
1721117211
"input_cost_per_token": 1.25e-06,
1721217212
"litellm_provider": "openrouter",
17213-
"max_input_tokens": 400000,
17213+
"max_input_tokens": 272000,
1721417214
"max_output_tokens": 128000,
1721517215
"max_tokens": 128000,
1721617216
"mode": "chat",
@@ -17229,7 +17229,7 @@
1722917229
"cache_read_input_token_cost": 1.25e-07,
1723017230
"input_cost_per_token": 1.25e-06,
1723117231
"litellm_provider": "openrouter",
17232-
"max_input_tokens": 400000,
17232+
"max_input_tokens": 272000,
1723317233
"max_output_tokens": 128000,
1723417234
"max_tokens": 128000,
1723517235
"mode": "chat",
@@ -17248,7 +17248,7 @@
1724817248
"cache_read_input_token_cost": 1.25e-07,
1724917249
"input_cost_per_token": 1.25e-06,
1725017250
"litellm_provider": "openrouter",
17251-
"max_input_tokens": 400000,
17251+
"max_input_tokens": 272000,
1725217252
"max_output_tokens": 128000,
1725317253
"max_tokens": 128000,
1725417254
"mode": "chat",
@@ -17267,7 +17267,7 @@
1726717267
"cache_read_input_token_cost": 2.5e-08,
1726817268
"input_cost_per_token": 2.5e-07,
1726917269
"litellm_provider": "openrouter",
17270-
"max_input_tokens": 400000,
17270+
"max_input_tokens": 272000,
1727117271
"max_output_tokens": 128000,
1727217272
"max_tokens": 128000,
1727317273
"mode": "chat",
@@ -17286,7 +17286,7 @@
1728617286
"cache_read_input_token_cost": 5e-09,
1728717287
"input_cost_per_token": 5e-08,
1728817288
"litellm_provider": "openrouter",
17289-
"max_input_tokens": 400000,
17289+
"max_input_tokens": 272000,
1729017290
"max_output_tokens": 128000,
1729117291
"max_tokens": 128000,
1729217292
"mode": "chat",

0 commit comments

Comments
 (0)