Skip to content

Commit 937b6eb

Browse files
committed
chore: re-generate all
1 parent bbb2aa2 commit 937b6eb

File tree

11 files changed

+898
-1686
lines changed

11 files changed

+898
-1686
lines changed

internal/providers/configs/aihubmix.json

Lines changed: 193 additions & 386 deletions
Large diffs are not rendered by default.

internal/providers/configs/avian.json

Lines changed: 4 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -23,8 +23,7 @@
2323
"high"
2424
],
2525
"default_reasoning_effort": "medium",
26-
"supports_attachments": false,
27-
"options": {}
26+
"supports_attachments": false
2827
},
2928
{
3029
"id": "z-ai/glm-5",
@@ -36,8 +35,7 @@
3635
"context_window": 204800,
3736
"default_max_tokens": 131072,
3837
"can_reason": false,
39-
"supports_attachments": false,
40-
"options": {}
38+
"supports_attachments": false
4139
},
4240
{
4341
"id": "moonshotai/kimi-k2.5",
@@ -49,8 +47,7 @@
4947
"context_window": 262144,
5048
"default_max_tokens": 262144,
5149
"can_reason": false,
52-
"supports_attachments": false,
53-
"options": {}
50+
"supports_attachments": false
5451
},
5552
{
5653
"id": "minimax/minimax-m2.5",
@@ -62,8 +59,7 @@
6259
"context_window": 196608,
6360
"default_max_tokens": 131072,
6461
"can_reason": false,
65-
"supports_attachments": false,
66-
"options": {}
62+
"supports_attachments": false
6763
}
6864
]
6965
}

internal/providers/configs/copilot.json

Lines changed: 25 additions & 50 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,7 @@
1616
"context_window": 200000,
1717
"default_max_tokens": 32000,
1818
"can_reason": false,
19-
"supports_attachments": true,
20-
"options": {}
19+
"supports_attachments": true
2120
},
2221
{
2322
"id": "claude-opus-4.5",
@@ -29,8 +28,7 @@
2928
"context_window": 200000,
3029
"default_max_tokens": 32000,
3130
"can_reason": false,
32-
"supports_attachments": true,
33-
"options": {}
31+
"supports_attachments": true
3432
},
3533
{
3634
"id": "claude-opus-4.6",
@@ -42,8 +40,7 @@
4240
"context_window": 200000,
4341
"default_max_tokens": 64000,
4442
"can_reason": false,
45-
"supports_attachments": true,
46-
"options": {}
43+
"supports_attachments": true
4744
},
4845
{
4946
"id": "claude-opus-4.6-fast",
@@ -55,8 +52,7 @@
5552
"context_window": 200000,
5653
"default_max_tokens": 64000,
5754
"can_reason": false,
58-
"supports_attachments": true,
59-
"options": {}
55+
"supports_attachments": true
6056
},
6157
{
6258
"id": "claude-sonnet-4",
@@ -68,8 +64,7 @@
6864
"context_window": 216000,
6965
"default_max_tokens": 16000,
7066
"can_reason": false,
71-
"supports_attachments": true,
72-
"options": {}
67+
"supports_attachments": true
7368
},
7469
{
7570
"id": "claude-sonnet-4.5",
@@ -81,8 +76,7 @@
8176
"context_window": 200000,
8277
"default_max_tokens": 32000,
8378
"can_reason": false,
84-
"supports_attachments": true,
85-
"options": {}
79+
"supports_attachments": true
8680
},
8781
{
8882
"id": "claude-sonnet-4.6",
@@ -94,8 +88,7 @@
9488
"context_window": 200000,
9589
"default_max_tokens": 32000,
9690
"can_reason": false,
97-
"supports_attachments": true,
98-
"options": {}
91+
"supports_attachments": true
9992
},
10093
{
10194
"id": "gemini-2.5-pro",
@@ -107,8 +100,7 @@
107100
"context_window": 128000,
108101
"default_max_tokens": 64000,
109102
"can_reason": false,
110-
"supports_attachments": true,
111-
"options": {}
103+
"supports_attachments": true
112104
},
113105
{
114106
"id": "gemini-3-flash-preview",
@@ -120,8 +112,7 @@
120112
"context_window": 128000,
121113
"default_max_tokens": 64000,
122114
"can_reason": false,
123-
"supports_attachments": true,
124-
"options": {}
115+
"supports_attachments": true
125116
},
126117
{
127118
"id": "gemini-3.1-pro-preview",
@@ -133,8 +124,7 @@
133124
"context_window": 128000,
134125
"default_max_tokens": 64000,
135126
"can_reason": false,
136-
"supports_attachments": true,
137-
"options": {}
127+
"supports_attachments": true
138128
},
139129
{
140130
"id": "goldeneye-free-auto",
@@ -146,8 +136,7 @@
146136
"context_window": 400000,
147137
"default_max_tokens": 128000,
148138
"can_reason": false,
149-
"supports_attachments": true,
150-
"options": {}
139+
"supports_attachments": true
151140
},
152141
{
153142
"id": "gpt-3.5-turbo-0613",
@@ -159,8 +148,7 @@
159148
"context_window": 16384,
160149
"default_max_tokens": 4096,
161150
"can_reason": false,
162-
"supports_attachments": false,
163-
"options": {}
151+
"supports_attachments": false
164152
},
165153
{
166154
"id": "gpt-4-0125-preview",
@@ -172,8 +160,7 @@
172160
"context_window": 128000,
173161
"default_max_tokens": 4096,
174162
"can_reason": false,
175-
"supports_attachments": false,
176-
"options": {}
163+
"supports_attachments": false
177164
},
178165
{
179166
"id": "gpt-4-0613",
@@ -185,8 +172,7 @@
185172
"context_window": 32768,
186173
"default_max_tokens": 4096,
187174
"can_reason": false,
188-
"supports_attachments": false,
189-
"options": {}
175+
"supports_attachments": false
190176
},
191177
{
192178
"id": "gpt-5-mini",
@@ -198,8 +184,7 @@
198184
"context_window": 264000,
199185
"default_max_tokens": 64000,
200186
"can_reason": false,
201-
"supports_attachments": true,
202-
"options": {}
187+
"supports_attachments": true
203188
},
204189
{
205190
"id": "gpt-5.1",
@@ -211,8 +196,7 @@
211196
"context_window": 264000,
212197
"default_max_tokens": 64000,
213198
"can_reason": false,
214-
"supports_attachments": true,
215-
"options": {}
199+
"supports_attachments": true
216200
},
217201
{
218202
"id": "gpt-5.1-codex",
@@ -224,8 +208,7 @@
224208
"context_window": 400000,
225209
"default_max_tokens": 128000,
226210
"can_reason": false,
227-
"supports_attachments": true,
228-
"options": {}
211+
"supports_attachments": true
229212
},
230213
{
231214
"id": "gpt-5.1-codex-max",
@@ -237,8 +220,7 @@
237220
"context_window": 400000,
238221
"default_max_tokens": 128000,
239222
"can_reason": false,
240-
"supports_attachments": true,
241-
"options": {}
223+
"supports_attachments": true
242224
},
243225
{
244226
"id": "gpt-5.1-codex-mini",
@@ -250,8 +232,7 @@
250232
"context_window": 400000,
251233
"default_max_tokens": 128000,
252234
"can_reason": false,
253-
"supports_attachments": true,
254-
"options": {}
235+
"supports_attachments": true
255236
},
256237
{
257238
"id": "gpt-5.2",
@@ -263,8 +244,7 @@
263244
"context_window": 264000,
264245
"default_max_tokens": 64000,
265246
"can_reason": false,
266-
"supports_attachments": true,
267-
"options": {}
247+
"supports_attachments": true
268248
},
269249
{
270250
"id": "gpt-5.2-codex",
@@ -276,8 +256,7 @@
276256
"context_window": 400000,
277257
"default_max_tokens": 128000,
278258
"can_reason": false,
279-
"supports_attachments": true,
280-
"options": {}
259+
"supports_attachments": true
281260
},
282261
{
283262
"id": "gpt-5.3-codex",
@@ -289,8 +268,7 @@
289268
"context_window": 400000,
290269
"default_max_tokens": 128000,
291270
"can_reason": false,
292-
"supports_attachments": true,
293-
"options": {}
271+
"supports_attachments": true
294272
},
295273
{
296274
"id": "gpt-5.4",
@@ -302,8 +280,7 @@
302280
"context_window": 400000,
303281
"default_max_tokens": 128000,
304282
"can_reason": false,
305-
"supports_attachments": true,
306-
"options": {}
283+
"supports_attachments": true
307284
},
308285
{
309286
"id": "gpt-5.4-mini",
@@ -315,8 +292,7 @@
315292
"context_window": 400000,
316293
"default_max_tokens": 128000,
317294
"can_reason": false,
318-
"supports_attachments": true,
319-
"options": {}
295+
"supports_attachments": true
320296
},
321297
{
322298
"id": "grok-code-fast-1",
@@ -328,8 +304,7 @@
328304
"context_window": 128000,
329305
"default_max_tokens": 64000,
330306
"can_reason": false,
331-
"supports_attachments": false,
332-
"options": {}
307+
"supports_attachments": false
333308
}
334309
]
335310
}

0 commit comments

Comments
 (0)