|
129 | 129 | "name": "Anthropic: Claude 3.5 Sonnet", |
130 | 130 | "cost_per_1m_in": 3, |
131 | 131 | "cost_per_1m_out": 15, |
132 | | - "cost_per_1m_in_cached": 3.75, |
133 | | - "cost_per_1m_out_cached": 0.3, |
| 132 | + "cost_per_1m_in_cached": 0, |
| 133 | + "cost_per_1m_out_cached": 0, |
134 | 134 | "context_window": 200000, |
135 | 135 | "default_max_tokens": 4096, |
136 | 136 | "can_reason": false, |
|
220 | 220 | "name": "Anthropic: Claude Sonnet 4.5", |
221 | 221 | "cost_per_1m_in": 3, |
222 | 222 | "cost_per_1m_out": 15, |
223 | | - "cost_per_1m_in_cached": 0, |
224 | | - "cost_per_1m_out_cached": 0, |
| 223 | + "cost_per_1m_in_cached": 3.75, |
| 224 | + "cost_per_1m_out_cached": 0.3, |
225 | 225 | "context_window": 1000000, |
226 | 226 | "default_max_tokens": 32000, |
227 | 227 | "can_reason": true, |
|
335 | 335 | { |
336 | 336 | "id": "deepseek/deepseek-v3.1-terminus", |
337 | 337 | "name": "DeepSeek: DeepSeek V3.1 Terminus", |
338 | | - "cost_per_1m_in": 0.22999999999999998, |
339 | | - "cost_per_1m_out": 0.8999999999999999, |
| 338 | + "cost_per_1m_in": 0.44999999999999996, |
| 339 | + "cost_per_1m_out": 1.5, |
340 | 340 | "cost_per_1m_in_cached": 0, |
341 | 341 | "cost_per_1m_out_cached": 0, |
342 | | - "context_window": 163840, |
343 | | - "default_max_tokens": 81920, |
| 342 | + "context_window": 131072, |
| 343 | + "default_max_tokens": 65536, |
344 | 344 | "can_reason": true, |
345 | 345 | "has_reasoning_efforts": true, |
346 | 346 | "supports_attachments": false |
347 | 347 | }, |
348 | 348 | { |
349 | 349 | "id": "deepseek/deepseek-v3.2-exp", |
350 | 350 | "name": "DeepSeek: DeepSeek V3.2 Exp", |
351 | | - "cost_per_1m_in": 0.27, |
| 351 | + "cost_per_1m_in": 0.28, |
352 | 352 | "cost_per_1m_out": 0.39999999999999997, |
353 | 353 | "cost_per_1m_in_cached": 0, |
354 | 354 | "cost_per_1m_out_cached": 0, |
355 | 355 | "context_window": 163840, |
356 | | - "default_max_tokens": 16384, |
| 356 | + "default_max_tokens": 81920, |
357 | 357 | "can_reason": true, |
358 | 358 | "has_reasoning_efforts": true, |
359 | 359 | "supports_attachments": false |
|
470 | 470 | "cost_per_1m_in_cached": 0, |
471 | 471 | "cost_per_1m_out_cached": 0, |
472 | 472 | "context_window": 1048576, |
473 | | - "default_max_tokens": 32768, |
| 473 | + "default_max_tokens": 32767, |
474 | 474 | "can_reason": true, |
475 | 475 | "has_reasoning_efforts": true, |
476 | 476 | "supports_attachments": true |
|
595 | 595 | { |
596 | 596 | "id": "meta-llama/llama-3.1-405b-instruct", |
597 | 597 | "name": "Meta: Llama 3.1 405B Instruct", |
598 | | - "cost_per_1m_in": 0.7999999999999999, |
599 | | - "cost_per_1m_out": 0.7999999999999999, |
| 598 | + "cost_per_1m_in": 3, |
| 599 | + "cost_per_1m_out": 3, |
600 | 600 | "cost_per_1m_in_cached": 0, |
601 | 601 | "cost_per_1m_out_cached": 0, |
602 | | - "context_window": 32768, |
603 | | - "default_max_tokens": 8192, |
| 602 | + "context_window": 131072, |
| 603 | + "default_max_tokens": 13107, |
604 | 604 | "can_reason": false, |
605 | 605 | "has_reasoning_efforts": false, |
606 | 606 | "supports_attachments": false |
|
621 | 621 | { |
622 | 622 | "id": "meta-llama/llama-3.1-8b-instruct", |
623 | 623 | "name": "Meta: Llama 3.1 8B Instruct", |
624 | | - "cost_per_1m_in": 0.02, |
625 | | - "cost_per_1m_out": 0.03, |
| 624 | + "cost_per_1m_in": 0.03, |
| 625 | + "cost_per_1m_out": 0.049999999999999996, |
626 | 626 | "cost_per_1m_in_cached": 0, |
627 | 627 | "cost_per_1m_out_cached": 0, |
628 | 628 | "context_window": 131072, |
|
648 | 648 | "id": "meta-llama/llama-3.3-70b-instruct", |
649 | 649 | "name": "Meta: Llama 3.3 70B Instruct", |
650 | 650 | "cost_per_1m_in": 0.13, |
651 | | - "cost_per_1m_out": 0.39, |
| 651 | + "cost_per_1m_out": 0.39999999999999997, |
652 | 652 | "cost_per_1m_in_cached": 0, |
653 | 653 | "cost_per_1m_out_cached": 0, |
654 | 654 | "context_window": 131072, |
655 | | - "default_max_tokens": 8192, |
| 655 | + "default_max_tokens": 65536, |
656 | 656 | "can_reason": false, |
657 | 657 | "has_reasoning_efforts": false, |
658 | 658 | "supports_attachments": false |
|
664 | 664 | "cost_per_1m_out": 0, |
665 | 665 | "cost_per_1m_in_cached": 0, |
666 | 666 | "cost_per_1m_out_cached": 0, |
667 | | - "context_window": 128000, |
668 | | - "default_max_tokens": 2014, |
| 667 | + "context_window": 65536, |
| 668 | + "default_max_tokens": 6553, |
669 | 669 | "can_reason": false, |
670 | 670 | "has_reasoning_efforts": false, |
671 | 671 | "supports_attachments": false |
|
1076 | 1076 | { |
1077 | 1077 | "id": "mistralai/mistral-small-3.2-24b-instruct", |
1078 | 1078 | "name": "Mistral: Mistral Small 3.2 24B", |
1079 | | - "cost_per_1m_in": 0.06, |
1080 | | - "cost_per_1m_out": 0.18, |
| 1079 | + "cost_per_1m_in": 0.09999999999999999, |
| 1080 | + "cost_per_1m_out": 0.3, |
1081 | 1081 | "cost_per_1m_in_cached": 0, |
1082 | 1082 | "cost_per_1m_out_cached": 0, |
1083 | 1083 | "context_window": 131072, |
1084 | | - "default_max_tokens": 65536, |
| 1084 | + "default_max_tokens": 13107, |
1085 | 1085 | "can_reason": false, |
1086 | 1086 | "has_reasoning_efforts": false, |
1087 | 1087 | "supports_attachments": true |
|
1167 | 1167 | { |
1168 | 1168 | "id": "moonshotai/kimi-k2-0905", |
1169 | 1169 | "name": "MoonshotAI: Kimi K2 0905", |
1170 | | - "cost_per_1m_in": 0.5, |
1171 | | - "cost_per_1m_out": 2, |
| 1170 | + "cost_per_1m_in": 1, |
| 1171 | + "cost_per_1m_out": 3, |
1172 | 1172 | "cost_per_1m_in_cached": 0, |
1173 | 1173 | "cost_per_1m_out_cached": 0, |
1174 | 1174 | "context_window": 262144, |
|
1547 | 1547 | "cost_per_1m_in": 0.049999999999999996, |
1548 | 1548 | "cost_per_1m_out": 0.39999999999999997, |
1549 | 1549 | "cost_per_1m_in_cached": 0, |
1550 | | - "cost_per_1m_out_cached": 0.005, |
| 1550 | + "cost_per_1m_out_cached": 0.01, |
1551 | 1551 | "context_window": 400000, |
1552 | | - "default_max_tokens": 64000, |
| 1552 | + "default_max_tokens": 40000, |
1553 | 1553 | "can_reason": true, |
1554 | 1554 | "has_reasoning_efforts": true, |
1555 | 1555 | "supports_attachments": true |
|
1583 | 1583 | { |
1584 | 1584 | "id": "openai/gpt-oss-20b", |
1585 | 1585 | "name": "OpenAI: gpt-oss-20b", |
1586 | | - "cost_per_1m_in": 0.03, |
1587 | | - "cost_per_1m_out": 0.14, |
| 1586 | + "cost_per_1m_in": 0.049999999999999996, |
| 1587 | + "cost_per_1m_out": 0.19999999999999998, |
1588 | 1588 | "cost_per_1m_in_cached": 0, |
1589 | 1589 | "cost_per_1m_out_cached": 0, |
1590 | 1590 | "context_window": 131072, |
1591 | | - "default_max_tokens": 13107, |
| 1591 | + "default_max_tokens": 65536, |
1592 | 1592 | "can_reason": true, |
1593 | 1593 | "has_reasoning_efforts": true, |
1594 | 1594 | "supports_attachments": false |
|
1697 | 1697 | "has_reasoning_efforts": false, |
1698 | 1698 | "supports_attachments": false |
1699 | 1699 | }, |
| 1700 | + { |
| 1701 | + "id": "qwen/qwq-32b", |
| 1702 | + "name": "Qwen: QwQ 32B", |
| 1703 | + "cost_per_1m_in": 0.15, |
| 1704 | + "cost_per_1m_out": 0.58, |
| 1705 | + "cost_per_1m_in_cached": 0, |
| 1706 | + "cost_per_1m_out_cached": 0, |
| 1707 | + "context_window": 131072, |
| 1708 | + "default_max_tokens": 65536, |
| 1709 | + "can_reason": true, |
| 1710 | + "has_reasoning_efforts": true, |
| 1711 | + "supports_attachments": false |
| 1712 | + }, |
1700 | 1713 | { |
1701 | 1714 | "id": "qwen/qwen-plus-2025-07-28", |
1702 | 1715 | "name": "Qwen: Qwen Plus 0728", |
|
1830 | 1843 | { |
1831 | 1844 | "id": "qwen/qwen3-30b-a3b-instruct-2507", |
1832 | 1845 | "name": "Qwen: Qwen3 30B A3B Instruct 2507", |
1833 | | - "cost_per_1m_in": 0.08, |
1834 | | - "cost_per_1m_out": 0.33, |
| 1846 | + "cost_per_1m_in": 0.09999999999999999, |
| 1847 | + "cost_per_1m_out": 0.3, |
1835 | 1848 | "cost_per_1m_in_cached": 0, |
1836 | 1849 | "cost_per_1m_out_cached": 0, |
1837 | 1850 | "context_window": 262144, |
1838 | | - "default_max_tokens": 131072, |
| 1851 | + "default_max_tokens": 26214, |
1839 | 1852 | "can_reason": false, |
1840 | 1853 | "has_reasoning_efforts": false, |
1841 | 1854 | "supports_attachments": false |
|
1882 | 1895 | { |
1883 | 1896 | "id": "qwen/qwen3-coder-30b-a3b-instruct", |
1884 | 1897 | "name": "Qwen: Qwen3 Coder 30B A3B Instruct", |
1885 | | - "cost_per_1m_in": 0.07, |
1886 | | - "cost_per_1m_out": 0.27, |
| 1898 | + "cost_per_1m_in": 0.09999999999999999, |
| 1899 | + "cost_per_1m_out": 0.3, |
1887 | 1900 | "cost_per_1m_in_cached": 0, |
1888 | 1901 | "cost_per_1m_out_cached": 0, |
1889 | 1902 | "context_window": 262144, |
|
2009 | 2022 | "has_reasoning_efforts": true, |
2010 | 2023 | "supports_attachments": true |
2011 | 2024 | }, |
| 2025 | + { |
| 2026 | + "id": "qwen/qwen3-vl-30b-a3b-instruct", |
| 2027 | + "name": "Qwen: Qwen3 VL 30B A3B Instruct", |
| 2028 | + "cost_per_1m_in": 0.3, |
| 2029 | + "cost_per_1m_out": 1, |
| 2030 | + "cost_per_1m_in_cached": 0, |
| 2031 | + "cost_per_1m_out_cached": 0, |
| 2032 | + "context_window": 262144, |
| 2033 | + "default_max_tokens": 131072, |
| 2034 | + "can_reason": false, |
| 2035 | + "has_reasoning_efforts": false, |
| 2036 | + "supports_attachments": true |
| 2037 | + }, |
| 2038 | + { |
| 2039 | + "id": "qwen/qwen3-vl-30b-a3b-thinking", |
| 2040 | + "name": "Qwen: Qwen3 VL 30B A3B Thinking", |
| 2041 | + "cost_per_1m_in": 0.3, |
| 2042 | + "cost_per_1m_out": 1, |
| 2043 | + "cost_per_1m_in_cached": 0, |
| 2044 | + "cost_per_1m_out_cached": 0, |
| 2045 | + "context_window": 262144, |
| 2046 | + "default_max_tokens": 131072, |
| 2047 | + "can_reason": true, |
| 2048 | + "has_reasoning_efforts": true, |
| 2049 | + "supports_attachments": true |
| 2050 | + }, |
2012 | 2051 | { |
2013 | 2052 | "id": "stepfun-ai/step3", |
2014 | 2053 | "name": "StepFun: Step3", |
|
2090 | 2129 | { |
2091 | 2130 | "id": "z-ai/glm-4.5", |
2092 | 2131 | "name": "Z.AI: GLM 4.5", |
2093 | | - "cost_per_1m_in": 0.35, |
2094 | | - "cost_per_1m_out": 1.55, |
| 2132 | + "cost_per_1m_in": 0.6, |
| 2133 | + "cost_per_1m_out": 2.2, |
2095 | 2134 | "cost_per_1m_in_cached": 0, |
2096 | | - "cost_per_1m_out_cached": 0, |
| 2135 | + "cost_per_1m_out_cached": 0.11, |
2097 | 2136 | "context_window": 131072, |
2098 | | - "default_max_tokens": 65536, |
| 2137 | + "default_max_tokens": 48000, |
2099 | 2138 | "can_reason": true, |
2100 | 2139 | "has_reasoning_efforts": true, |
2101 | 2140 | "supports_attachments": false |
|
2155 | 2194 | { |
2156 | 2195 | "id": "x-ai/grok-3", |
2157 | 2196 | "name": "xAI: Grok 3", |
2158 | | - "cost_per_1m_in": 5, |
2159 | | - "cost_per_1m_out": 25, |
| 2197 | + "cost_per_1m_in": 3, |
| 2198 | + "cost_per_1m_out": 15, |
2160 | 2199 | "cost_per_1m_in_cached": 0, |
2161 | | - "cost_per_1m_out_cached": 1.25, |
| 2200 | + "cost_per_1m_out_cached": 0.75, |
2162 | 2201 | "context_window": 131072, |
2163 | 2202 | "default_max_tokens": 13107, |
2164 | 2203 | "can_reason": false, |
|
2168 | 2207 | { |
2169 | 2208 | "id": "x-ai/grok-3-beta", |
2170 | 2209 | "name": "xAI: Grok 3 Beta", |
2171 | | - "cost_per_1m_in": 5, |
2172 | | - "cost_per_1m_out": 25, |
| 2210 | + "cost_per_1m_in": 3, |
| 2211 | + "cost_per_1m_out": 15, |
2173 | 2212 | "cost_per_1m_in_cached": 0, |
2174 | | - "cost_per_1m_out_cached": 1.25, |
| 2213 | + "cost_per_1m_out_cached": 0.75, |
2175 | 2214 | "context_window": 131072, |
2176 | 2215 | "default_max_tokens": 13107, |
2177 | 2216 | "can_reason": false, |
|
2181 | 2220 | { |
2182 | 2221 | "id": "x-ai/grok-3-mini", |
2183 | 2222 | "name": "xAI: Grok 3 Mini", |
2184 | | - "cost_per_1m_in": 0.6, |
2185 | | - "cost_per_1m_out": 4, |
| 2223 | + "cost_per_1m_in": 0.3, |
| 2224 | + "cost_per_1m_out": 0.5, |
2186 | 2225 | "cost_per_1m_in_cached": 0, |
2187 | | - "cost_per_1m_out_cached": 0.15, |
| 2226 | + "cost_per_1m_out_cached": 0.075, |
2188 | 2227 | "context_window": 131072, |
2189 | 2228 | "default_max_tokens": 13107, |
2190 | 2229 | "can_reason": true, |
|
2194 | 2233 | { |
2195 | 2234 | "id": "x-ai/grok-3-mini-beta", |
2196 | 2235 | "name": "xAI: Grok 3 Mini Beta", |
2197 | | - "cost_per_1m_in": 0.6, |
2198 | | - "cost_per_1m_out": 4, |
| 2236 | + "cost_per_1m_in": 0.3, |
| 2237 | + "cost_per_1m_out": 0.5, |
2199 | 2238 | "cost_per_1m_in_cached": 0, |
2200 | | - "cost_per_1m_out_cached": 0.15, |
| 2239 | + "cost_per_1m_out_cached": 0.075, |
2201 | 2240 | "context_window": 131072, |
2202 | 2241 | "default_max_tokens": 13107, |
2203 | 2242 | "can_reason": true, |
|
0 commit comments