|
10 | 10 | { |
11 | 11 | "id": "qwen/qwen3-next-80b-a3b-thinking", |
12 | 12 | "name": "Qwen: Qwen3 Next 80B A3B Thinking", |
13 | | - "cost_per_1m_in": 0.14673906, |
14 | | - "cost_per_1m_out": 0.586956456, |
| 13 | + "cost_per_1m_in": 0.14, |
| 14 | + "cost_per_1m_out": 1.4, |
15 | 15 | "cost_per_1m_in_cached": 0, |
16 | 16 | "cost_per_1m_out_cached": 0, |
17 | 17 | "context_window": 262144, |
|
23 | 23 | { |
24 | 24 | "id": "qwen/qwen3-next-80b-a3b-instruct", |
25 | 25 | "name": "Qwen: Qwen3 Next 80B A3B Instruct", |
26 | | - "cost_per_1m_in": 0.3, |
27 | | - "cost_per_1m_out": 0.3, |
| 26 | + "cost_per_1m_in": 0.14, |
| 27 | + "cost_per_1m_out": 1.4, |
28 | 28 | "cost_per_1m_in_cached": 0, |
29 | 29 | "cost_per_1m_out_cached": 0, |
30 | 30 | "context_window": 262144, |
31 | | - "default_max_tokens": 131072, |
32 | | - "can_reason": false, |
33 | | - "has_reasoning_efforts": false, |
34 | | - "supports_attachments": false |
35 | | - }, |
36 | | - { |
37 | | - "id": "meituan/longcat-flash-chat", |
38 | | - "name": "Meituan: LongCat Flash Chat", |
39 | | - "cost_per_1m_in": 0.24999987999999998, |
40 | | - "cost_per_1m_out": 0.999999888, |
41 | | - "cost_per_1m_in_cached": 0, |
42 | | - "cost_per_1m_out_cached": 0, |
43 | | - "context_window": 131072, |
44 | | - "default_max_tokens": 13107, |
| 31 | + "default_max_tokens": 26214, |
45 | 32 | "can_reason": false, |
46 | 33 | "has_reasoning_efforts": false, |
47 | 34 | "supports_attachments": false |
|
140 | 127 | { |
141 | 128 | "id": "moonshotai/kimi-k2-0905", |
142 | 129 | "name": "MoonshotAI: Kimi K2 0905", |
143 | | - "cost_per_1m_in": 0.58, |
144 | | - "cost_per_1m_out": 2.29, |
| 130 | + "cost_per_1m_in": 0.5, |
| 131 | + "cost_per_1m_out": 2, |
145 | 132 | "cost_per_1m_in_cached": 0, |
146 | 133 | "cost_per_1m_out_cached": 0, |
147 | 134 | "context_window": 262144, |
148 | | - "default_max_tokens": 131072, |
| 135 | + "default_max_tokens": 26214, |
149 | 136 | "can_reason": false, |
150 | 137 | "has_reasoning_efforts": false, |
151 | 138 | "supports_attachments": false |
|
244 | 231 | { |
245 | 232 | "id": "deepseek/deepseek-chat-v3.1", |
246 | 233 | "name": "DeepSeek: DeepSeek V3.1", |
247 | | - "cost_per_1m_in": 0.27, |
248 | | - "cost_per_1m_out": 1, |
| 234 | + "cost_per_1m_in": 0.56, |
| 235 | + "cost_per_1m_out": 1.68, |
249 | 236 | "cost_per_1m_in_cached": 0, |
250 | 237 | "cost_per_1m_out_cached": 0, |
251 | 238 | "context_window": 163840, |
|
283 | 270 | { |
284 | 271 | "id": "z-ai/glm-4.5v", |
285 | 272 | "name": "Z.AI: GLM 4.5V", |
286 | | - "cost_per_1m_in": 0.5, |
| 273 | + "cost_per_1m_in": 0.6, |
287 | 274 | "cost_per_1m_out": 1.7999999999999998, |
288 | 275 | "cost_per_1m_in_cached": 0, |
289 | | - "cost_per_1m_out_cached": 0, |
| 276 | + "cost_per_1m_out_cached": 0.11, |
290 | 277 | "context_window": 65536, |
291 | | - "default_max_tokens": 32768, |
| 278 | + "default_max_tokens": 8192, |
292 | 279 | "can_reason": true, |
293 | 280 | "has_reasoning_efforts": true, |
294 | 281 | "supports_attachments": true |
|
361 | 348 | { |
362 | 349 | "id": "openai/gpt-oss-120b", |
363 | 350 | "name": "OpenAI: gpt-oss-120b", |
364 | | - "cost_per_1m_in": 0.15, |
365 | | - "cost_per_1m_out": 0.6, |
| 351 | + "cost_per_1m_in": 0.09999999999999999, |
| 352 | + "cost_per_1m_out": 0.49, |
366 | 353 | "cost_per_1m_in_cached": 0, |
367 | 354 | "cost_per_1m_out_cached": 0, |
368 | 355 | "context_window": 131072, |
369 | | - "default_max_tokens": 16384, |
| 356 | + "default_max_tokens": 13107, |
370 | 357 | "can_reason": true, |
371 | 358 | "has_reasoning_efforts": true, |
372 | 359 | "supports_attachments": false |
|
439 | 426 | { |
440 | 427 | "id": "z-ai/glm-4.5", |
441 | 428 | "name": "Z.AI: GLM 4.5", |
442 | | - "cost_per_1m_in": 0.6, |
443 | | - "cost_per_1m_out": 2.2, |
| 429 | + "cost_per_1m_in": 0.41249980199999997, |
| 430 | + "cost_per_1m_out": 1.6499998152000002, |
444 | 431 | "cost_per_1m_in_cached": 0, |
445 | 432 | "cost_per_1m_out_cached": 0, |
446 | 433 | "context_window": 131072, |
|
517 | 504 | { |
518 | 505 | "id": "qwen/qwen3-coder", |
519 | 506 | "name": "Qwen: Qwen3 Coder 480B A35B", |
520 | | - "cost_per_1m_in": 0.39999999999999997, |
521 | | - "cost_per_1m_out": 1.7999999999999998, |
| 507 | + "cost_per_1m_in": 0.29, |
| 508 | + "cost_per_1m_out": 1.2, |
522 | 509 | "cost_per_1m_in_cached": 0, |
523 | 510 | "cost_per_1m_out_cached": 0, |
524 | 511 | "context_window": 262144, |
525 | | - "default_max_tokens": 131072, |
| 512 | + "default_max_tokens": 32768, |
526 | 513 | "can_reason": false, |
527 | 514 | "has_reasoning_efforts": false, |
528 | 515 | "supports_attachments": false |
|
543 | 530 | { |
544 | 531 | "id": "qwen/qwen3-235b-a22b-2507", |
545 | 532 | "name": "Qwen: Qwen3 235B A22B Instruct 2507", |
546 | | - "cost_per_1m_in": 0.22, |
547 | | - "cost_per_1m_out": 0.7999999999999999, |
| 533 | + "cost_per_1m_in": 0.13, |
| 534 | + "cost_per_1m_out": 0.6, |
548 | 535 | "cost_per_1m_in_cached": 0, |
549 | 536 | "cost_per_1m_out_cached": 0, |
550 | 537 | "context_window": 262144, |
551 | | - "default_max_tokens": 131072, |
| 538 | + "default_max_tokens": 26214, |
552 | 539 | "can_reason": false, |
553 | 540 | "has_reasoning_efforts": false, |
554 | 541 | "supports_attachments": false |
|
644 | 631 | "has_reasoning_efforts": false, |
645 | 632 | "supports_attachments": true |
646 | 633 | }, |
647 | | - { |
648 | | - "id": "minimax/minimax-m1", |
649 | | - "name": "MiniMax: MiniMax M1", |
650 | | - "cost_per_1m_in": 0.55, |
651 | | - "cost_per_1m_out": 2.2, |
652 | | - "cost_per_1m_in_cached": 0, |
653 | | - "cost_per_1m_out_cached": 0, |
654 | | - "context_window": 1000000, |
655 | | - "default_max_tokens": 20000, |
656 | | - "can_reason": true, |
657 | | - "has_reasoning_efforts": true, |
658 | | - "supports_attachments": false |
659 | | - }, |
660 | 634 | { |
661 | 635 | "id": "google/gemini-2.5-flash-lite-preview-06-17", |
662 | 636 | "name": "Google: Gemini 2.5 Flash Lite Preview 06-17", |
|
742 | 716 | "cost_per_1m_out": 1.5, |
743 | 717 | "cost_per_1m_in_cached": 0, |
744 | 718 | "cost_per_1m_out_cached": 0, |
745 | | - "context_window": 40960, |
| 719 | + "context_window": 40000, |
746 | 720 | "default_max_tokens": 20000, |
747 | 721 | "can_reason": true, |
748 | 722 | "has_reasoning_efforts": true, |
|
946 | 920 | { |
947 | 921 | "id": "qwen/qwen3-30b-a3b", |
948 | 922 | "name": "Qwen: Qwen3 30B A3B", |
949 | | - "cost_per_1m_in": 0.09, |
950 | | - "cost_per_1m_out": 0.44999999999999996, |
| 923 | + "cost_per_1m_in": 0.15, |
| 924 | + "cost_per_1m_out": 0.6, |
951 | 925 | "cost_per_1m_in_cached": 0, |
952 | 926 | "cost_per_1m_out_cached": 0, |
953 | 927 | "context_window": 131072, |
954 | | - "default_max_tokens": 65536, |
| 928 | + "default_max_tokens": 4000, |
955 | 929 | "can_reason": true, |
956 | 930 | "has_reasoning_efforts": true, |
957 | 931 | "supports_attachments": false |
|
982 | 956 | "has_reasoning_efforts": true, |
983 | 957 | "supports_attachments": false |
984 | 958 | }, |
985 | | - { |
986 | | - "id": "qwen/qwen3-235b-a22b:free", |
987 | | - "name": "Qwen: Qwen3 235B A22B (free)", |
988 | | - "cost_per_1m_in": 0, |
989 | | - "cost_per_1m_out": 0, |
990 | | - "cost_per_1m_in_cached": 0, |
991 | | - "cost_per_1m_out_cached": 0, |
992 | | - "context_window": 131072, |
993 | | - "default_max_tokens": 13107, |
994 | | - "can_reason": false, |
995 | | - "has_reasoning_efforts": false, |
996 | | - "supports_attachments": false |
997 | | - }, |
998 | 959 | { |
999 | 960 | "id": "qwen/qwen3-235b-a22b", |
1000 | 961 | "name": "Qwen: Qwen3 235B A22B", |
|
1128 | 1089 | { |
1129 | 1090 | "id": "meta-llama/llama-4-maverick", |
1130 | 1091 | "name": "Meta: Llama 4 Maverick", |
1131 | | - "cost_per_1m_in": 0.18, |
1132 | | - "cost_per_1m_out": 0.6, |
| 1092 | + "cost_per_1m_in": 0.22, |
| 1093 | + "cost_per_1m_out": 0.88, |
1133 | 1094 | "cost_per_1m_in_cached": 0, |
1134 | 1095 | "cost_per_1m_out_cached": 0, |
1135 | 1096 | "context_window": 1048576, |
1136 | | - "default_max_tokens": 524288, |
| 1097 | + "default_max_tokens": 104857, |
1137 | 1098 | "can_reason": false, |
1138 | 1099 | "has_reasoning_efforts": false, |
1139 | 1100 | "supports_attachments": true |
|
1180 | 1141 | { |
1181 | 1142 | "id": "deepseek/deepseek-chat-v3-0324", |
1182 | 1143 | "name": "DeepSeek: DeepSeek V3 0324", |
1183 | | - "cost_per_1m_in": 0.28, |
1184 | | - "cost_per_1m_out": 1.1400000000000001, |
| 1144 | + "cost_per_1m_in": 0.27, |
| 1145 | + "cost_per_1m_out": 1.13, |
1185 | 1146 | "cost_per_1m_in_cached": 0, |
1186 | 1147 | "cost_per_1m_out_cached": 0, |
1187 | 1148 | "context_window": 163840, |
|
1250 | 1211 | "cost_per_1m_in_cached": 3.75, |
1251 | 1212 | "cost_per_1m_out_cached": 0.3, |
1252 | 1213 | "context_window": 200000, |
1253 | | - "default_max_tokens": 64000, |
| 1214 | + "default_max_tokens": 32000, |
1254 | 1215 | "can_reason": true, |
1255 | 1216 | "has_reasoning_efforts": true, |
1256 | 1217 | "supports_attachments": true |
|
1457 | 1418 | "cost_per_1m_out": 0, |
1458 | 1419 | "cost_per_1m_in_cached": 0, |
1459 | 1420 | "cost_per_1m_out_cached": 0, |
1460 | | - "context_window": 65536, |
1461 | | - "default_max_tokens": 6553, |
| 1421 | + "context_window": 128000, |
| 1422 | + "default_max_tokens": 2014, |
1462 | 1423 | "can_reason": false, |
1463 | 1424 | "has_reasoning_efforts": false, |
1464 | 1425 | "supports_attachments": false |
1465 | 1426 | }, |
1466 | 1427 | { |
1467 | 1428 | "id": "meta-llama/llama-3.3-70b-instruct", |
1468 | 1429 | "name": "Meta: Llama 3.3 70B Instruct", |
1469 | | - "cost_per_1m_in": 0.039, |
1470 | | - "cost_per_1m_out": 0.12, |
| 1430 | + "cost_per_1m_in": 0.09999999999999999, |
| 1431 | + "cost_per_1m_out": 0.25, |
1471 | 1432 | "cost_per_1m_in_cached": 0, |
1472 | 1433 | "cost_per_1m_out_cached": 0, |
1473 | 1434 | "context_window": 131072, |
1474 | | - "default_max_tokens": 4096, |
| 1435 | + "default_max_tokens": 13107, |
1475 | 1436 | "can_reason": false, |
1476 | 1437 | "has_reasoning_efforts": false, |
1477 | 1438 | "supports_attachments": false |
|
1661 | 1622 | { |
1662 | 1623 | "id": "thedrummer/rocinante-12b", |
1663 | 1624 | "name": "TheDrummer: Rocinante 12B", |
1664 | | - "cost_per_1m_in": 0.16999999999999998, |
1665 | | - "cost_per_1m_out": 0.43, |
| 1625 | + "cost_per_1m_in": 0.24, |
| 1626 | + "cost_per_1m_out": 0.44999999999999996, |
1666 | 1627 | "cost_per_1m_in_cached": 0, |
1667 | 1628 | "cost_per_1m_out_cached": 0, |
1668 | 1629 | "context_window": 32768, |
1669 | | - "default_max_tokens": 3276, |
| 1630 | + "default_max_tokens": 16384, |
1670 | 1631 | "can_reason": false, |
1671 | 1632 | "has_reasoning_efforts": false, |
1672 | 1633 | "supports_attachments": false |
|
1736 | 1697 | "has_reasoning_efforts": false, |
1737 | 1698 | "supports_attachments": false |
1738 | 1699 | }, |
| 1700 | + { |
| 1701 | + "id": "nousresearch/hermes-3-llama-3.1-70b", |
| 1702 | + "name": "Nous: Hermes 3 70B Instruct", |
| 1703 | + "cost_per_1m_in": 0.39999999999999997, |
| 1704 | + "cost_per_1m_out": 0.39999999999999997, |
| 1705 | + "cost_per_1m_in_cached": 0, |
| 1706 | + "cost_per_1m_out_cached": 0, |
| 1707 | + "context_window": 12288, |
| 1708 | + "default_max_tokens": 1228, |
| 1709 | + "can_reason": false, |
| 1710 | + "has_reasoning_efforts": false, |
| 1711 | + "supports_attachments": false |
| 1712 | + }, |
1739 | 1713 | { |
1740 | 1714 | "id": "openai/gpt-4o-2024-08-06", |
1741 | 1715 | "name": "OpenAI: GPT-4o (2024-08-06)", |
|
1752 | 1726 | { |
1753 | 1727 | "id": "meta-llama/llama-3.1-8b-instruct", |
1754 | 1728 | "name": "Meta: Llama 3.1 8B Instruct", |
1755 | | - "cost_per_1m_in": 0.09999999999999999, |
1756 | | - "cost_per_1m_out": 0.09999999999999999, |
| 1729 | + "cost_per_1m_in": 0.03, |
| 1730 | + "cost_per_1m_out": 0.049999999999999996, |
1757 | 1731 | "cost_per_1m_in_cached": 0, |
1758 | 1732 | "cost_per_1m_out_cached": 0, |
1759 | 1733 | "context_window": 131072, |
1760 | | - "default_max_tokens": 4000, |
| 1734 | + "default_max_tokens": 8192, |
1761 | 1735 | "can_reason": false, |
1762 | 1736 | "has_reasoning_efforts": false, |
1763 | 1737 | "supports_attachments": false |
1764 | 1738 | }, |
1765 | 1739 | { |
1766 | 1740 | "id": "meta-llama/llama-3.1-405b-instruct", |
1767 | 1741 | "name": "Meta: Llama 3.1 405B Instruct", |
1768 | | - "cost_per_1m_in": 3, |
1769 | | - "cost_per_1m_out": 3, |
| 1742 | + "cost_per_1m_in": 0.7999999999999999, |
| 1743 | + "cost_per_1m_out": 0.7999999999999999, |
1770 | 1744 | "cost_per_1m_in_cached": 0, |
1771 | 1745 | "cost_per_1m_out_cached": 0, |
1772 | | - "context_window": 131072, |
1773 | | - "default_max_tokens": 13107, |
| 1746 | + "context_window": 32768, |
| 1747 | + "default_max_tokens": 8192, |
1774 | 1748 | "can_reason": false, |
1775 | 1749 | "has_reasoning_efforts": false, |
1776 | 1750 | "supports_attachments": false |
1777 | 1751 | }, |
1778 | 1752 | { |
1779 | 1753 | "id": "meta-llama/llama-3.1-70b-instruct", |
1780 | 1754 | "name": "Meta: Llama 3.1 70B Instruct", |
1781 | | - "cost_per_1m_in": 0.88, |
1782 | | - "cost_per_1m_out": 0.88, |
| 1755 | + "cost_per_1m_in": 0.09999999999999999, |
| 1756 | + "cost_per_1m_out": 0.28, |
1783 | 1757 | "cost_per_1m_in_cached": 0, |
1784 | 1758 | "cost_per_1m_out_cached": 0, |
1785 | 1759 | "context_window": 131072, |
1786 | | - "default_max_tokens": 13107, |
| 1760 | + "default_max_tokens": 8192, |
1787 | 1761 | "can_reason": false, |
1788 | 1762 | "has_reasoning_efforts": false, |
1789 | 1763 | "supports_attachments": false |
1790 | 1764 | }, |
1791 | 1765 | { |
1792 | 1766 | "id": "mistralai/mistral-nemo", |
1793 | 1767 | "name": "Mistral: Mistral Nemo", |
1794 | | - "cost_per_1m_in": 0.15, |
1795 | | - "cost_per_1m_out": 0.15, |
| 1768 | + "cost_per_1m_in": 0.08, |
| 1769 | + "cost_per_1m_out": 0.14, |
1796 | 1770 | "cost_per_1m_in_cached": 0, |
1797 | 1771 | "cost_per_1m_out_cached": 0, |
1798 | 1772 | "context_window": 131072, |
1799 | | - "default_max_tokens": 13107, |
| 1773 | + "default_max_tokens": 65536, |
1800 | 1774 | "can_reason": false, |
1801 | 1775 | "has_reasoning_efforts": false, |
1802 | 1776 | "supports_attachments": false |
|
1937 | 1911 | "cost_per_1m_in": 2.5, |
1938 | 1912 | "cost_per_1m_out": 10, |
1939 | 1913 | "cost_per_1m_in_cached": 0, |
1940 | | - "cost_per_1m_out_cached": 0, |
| 1914 | + "cost_per_1m_out_cached": 1.25, |
1941 | 1915 | "context_window": 128000, |
1942 | 1916 | "default_max_tokens": 8192, |
1943 | 1917 | "can_reason": false, |
|
0 commit comments