From 316ac298df3c87c322ee7f19e12c7b89d2f06144 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Sat, 21 Mar 2026 02:35:04 +0000 Subject: [PATCH] Update the models metadata --- langgraphics/metadata/models.json | 521 +++++++++++++++++------------- 1 file changed, 303 insertions(+), 218 deletions(-) diff --git a/langgraphics/metadata/models.json b/langgraphics/metadata/models.json index 55e7626..c8511d9 100644 --- a/langgraphics/metadata/models.json +++ b/langgraphics/metadata/models.json @@ -468,6 +468,22 @@ "cache_write": 0 } }, + "xiaomi/mimo-v2-omni": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 65536, + "context": 262144 + }, + "cost": { + "input": 0.4, + "output": 2, + "cache_read": 0, + "cache_write": 0 + } + }, "xiaomi/mimo-v2-flash-free": { "reasoning": true, "tool_call": true, @@ -501,6 +517,22 @@ "cache_write": 0 } }, + "xiaomi/mimo-v2-pro": { + "reasoning": true, + "tool_call": true, + "attachment": false, + "temperature": true, + "limit": { + "output": 128000, + "context": 1000000 + }, + "cost": { + "input": 1, + "output": 3, + "cache_read": 0.19999999999999998, + "cache_write": 0 + } + }, "kuaishou/kat-coder-pro-v1-free": { "reasoning": false, "tool_call": true, @@ -693,6 +725,22 @@ "cache_write": 0.0024 } }, + "volcengine/doubao-seed-2.0-code": { + "reasoning": false, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 32000, + "context": 256000 + }, + "cost": { + "input": 0.9, + "output": 4.48, + "cache_read": 0, + "cache_write": 0 + } + }, "deepseek/deepseek-v3.2": { "reasoning": false, "tool_call": true, @@ -824,6 +872,23 @@ "reasoning": 2.5 } }, + "google/gemini-3.1-flash-lite-preview": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 65536, + "context": 1048576 + }, + "cost": { + "input": 0.25, + "output": 1.5, + "cache_read": 0, + "cache_write": 0, + "reasoning": 1.5 + } + }, "google/gemini-3-flash-preview": { "reasoning": true, "tool_call": true, @@ -875,6 +940,23 @@ "reasoning": 12 } }, + "google/gemini-3-pro-image-preview": { + "reasoning": true, + "tool_call": false, + "attachment": true, + "temperature": true, + "limit": { + "output": 32768, + "context": 65536 + }, + "cost": { + "input": 2, + "output": 12, + "cache_read": 0, + "cache_write": 0, + "reasoning": 12 + } + }, "google/gemini-3-pro-preview": { "reasoning": true, "tool_call": true, @@ -1054,6 +1136,22 @@ "cache_write": 0 } }, + "z-ai/glm-5-turbo": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 128000, + "context": 200000 + }, + "cost": { + "input": 0.88, + "output": 3.48, + "cache_read": 0, + "cache_write": 0 + } + }, "z-ai/glm-4.6v": { "reasoning": true, "tool_call": true, @@ -1070,6 +1168,39 @@ "cache_write": 0 } }, + "qwen/qwen3.5-flash": { + "reasoning": false, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 1020000, + "context": 1020000 + }, + "cost": { + "input": 0.1, + "output": 0.4, + "cache_read": 0, + "cache_write": 0 + } + }, + "qwen/qwen3.5-plus": { + "reasoning": true, + "tool_call": true, + "attachment": false, + "temperature": true, + "limit": { + "output": 65536, + "context": 1000000 + }, + "cost": { + "input": 0.4, + "output": 2.4, + "cache_read": 0, + "cache_write": 0, + "reasoning": 2.4 + } + }, "qwen/qwen3-max": { "reasoning": false, "tool_call": true, @@ -1185,6 +1316,54 @@ "cache_write": 0 } }, + "x-ai/grok-4.2-fast": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 30000, + "context": 2000000 + }, + "cost": { + "input": 3, + "output": 9, + "cache_read": 0, + "cache_write": 0 + } + }, + "x-ai/grok-4.2-fast-non-reasoning": { + "reasoning": false, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 30000, + "context": 2000000 + }, + "cost": { + "input": 3, + "output": 9, + "cache_read": 0, + "cache_write": 0 + } + }, + "openai/gpt-5.3-codex": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": false, + "limit": { + "output": 128000, + "context": 400000 + }, + "cost": { + "input": 1.75, + "output": 14, + "cache_read": 0, + "cache_write": 0 + } + }, "openai/gpt-5-codex": { "reasoning": false, "tool_call": false, @@ -1304,6 +1483,54 @@ "cache_write": 0 } }, + "openai/gpt-5.4": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": false, + "limit": { + "output": 128000, + "context": 1050000 + }, + "cost": { + "input": 2.5, + "output": 15, + "cache_read": 0, + "cache_write": 0 + } + }, + "openai/gpt-5.4-pro": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": false, + "limit": { + "output": 128000, + "context": 1050000 + }, + "cost": { + "input": 30, + "output": 180, + "cache_read": 0, + "cache_write": 0 + } + }, + "openai/gpt-5.3-chat": { + "reasoning": false, + "tool_call": true, + "attachment": true, + "temperature": false, + "limit": { + "output": 16384, + "context": 128000 + }, + "cost": { + "input": 1.75, + "output": 14, + "cache_read": 0, + "cache_write": 0 + } + }, "openai/gpt-5.1-codex": { "reasoning": true, "tool_call": false, @@ -1338,6 +1565,40 @@ "cache_write": 0 } }, + "openai/gpt-5.4-nano": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 128000, + "context": 400000, + "input": 272000 + }, + "cost": { + "input": 0.19999999999999998, + "output": 1.25, + "cache_read": 0.02, + "cache_write": 0 + } + }, + "openai/gpt-5.4-mini": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 128000, + "context": 400000, + "input": 272000 + }, + "cost": { + "input": 0.75, + "output": 4.5, + "cache_read": 0.075, + "cache_write": 0 + } + }, "minimax/minimax-m2.5-lightning": { "reasoning": true, "tool_call": true, @@ -1371,6 +1632,38 @@ "cache_write": 0 } }, + "minimax/minimax-m2.7": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 131000, + "context": 204800 + }, + "cost": { + "input": 0.3, + "output": 1.2, + "cache_read": 0.06, + "cache_write": 0.375 + } + }, + "minimax/minimax-m2.7-highspeed": { + "reasoning": true, + "tool_call": true, + "attachment": true, + "temperature": true, + "limit": { + "output": 131100, + "context": 204800 + }, + "cost": { + "input": 0.6, + "output": 2.4, + "cache_read": 0.06, + "cache_write": 0.375 + } + }, "minimax/minimax-m2": { "reasoning": true, "tool_call": true, @@ -3768,22 +4061,6 @@ "cache_write": 0 } }, - "openai/gpt-5.4": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": false, - "limit": { - "output": 128000, - "context": 1050000 - }, - "cost": { - "input": 2.5, - "output": 15, - "cache_read": 0, - "cache_write": 0 - } - }, "perplexity/sonar": { "reasoning": false, "tool_call": false, @@ -7168,22 +7445,6 @@ "cache_write": 0 } }, - "openai/gpt-5.3-codex": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": false, - "limit": { - "output": 128000, - "context": 400000 - }, - "cost": { - "input": 1.75, - "output": 14, - "cache_read": 0, - "cache_write": 0 - } - }, "openai/gpt-4o-mini": { "reasoning": false, "tool_call": false, @@ -9898,22 +10159,6 @@ "cache_write": 0 } }, - "openai/gpt-5.4-pro": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": false, - "limit": { - "output": 128000, - "context": 1050000 - }, - "cost": { - "input": 30, - "output": 180, - "cache_read": 0, - "cache_write": 0 - } - }, "openai/o1-pro": { "reasoning": false, "tool_call": false, @@ -16688,38 +16933,6 @@ "cache_write": 0 } }, - "xiaomi/mimo-v2-omni": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 65536, - "context": 262144 - }, - "cost": { - "input": 0.4, - "output": 2, - "cache_read": 0, - "cache_write": 0 - } - }, - "xiaomi/mimo-v2-pro": { - "reasoning": true, - "tool_call": true, - "attachment": false, - "temperature": true, - "limit": { - "output": 65536, - "context": 1048576 - }, - "cost": { - "input": 1, - "output": 3, - "cache_read": 0, - "cache_write": 0 - } - }, "microsoft/mai-ds-r1:free": { "reasoning": true, "tool_call": true, @@ -17412,23 +17625,6 @@ "cache_write": 0.083333 } }, - "google/gemini-3.1-flash-lite-preview": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 65536, - "context": 1048576 - }, - "cost": { - "input": 0.25, - "output": 1.5, - "cache_read": 0, - "cache_write": 0, - "reasoning": 1.5 - } - }, "google/gemma-3-12b-it:free": { "reasoning": false, "tool_call": false, @@ -18379,23 +18575,6 @@ "cache_write": 0 } }, - "openai/gpt-5.4-nano": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 128000, - "context": 400000, - "input": 272000 - }, - "cost": { - "input": 0.19999999999999998, - "output": 1.25, - "cache_read": 0.02, - "cache_write": 0 - } - }, "openai/gpt-oss-120b:free": { "reasoning": true, "tool_call": true, @@ -18412,23 +18591,6 @@ "cache_write": 0 } }, - "openai/gpt-5.4-mini": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 128000, - "context": 400000, - "input": 272000 - }, - "cost": { - "input": 0.75, - "output": 4.5, - "cache_read": 0.075, - "cache_write": 0 - } - }, "minimax/minimax-m1": { "reasoning": true, "tool_call": true, @@ -18462,22 +18624,6 @@ "cache_write": 0 } }, - "minimax/minimax-m2.7": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 131000, - "context": 204800 - }, - "cost": { - "input": 0.3, - "output": 1.2, - "cache_read": 0.06, - "cache_write": 0.375 - } - }, "bytedance-seed/seedream-4.5": { "reasoning": false, "tool_call": false, @@ -19436,7 +19582,8 @@ "temperature": true, "limit": { "output": 32000, - "context": 128000 + "context": 200000, + "input": 128000 }, "cost": { "input": 0, @@ -19452,7 +19599,8 @@ "temperature": true, "limit": { "output": 32000, - "context": 128000 + "context": 144000, + "input": 128000 }, "cost": { "input": 0, @@ -19468,7 +19616,8 @@ "temperature": true, "limit": { "output": 32000, - "context": 128000 + "context": 160000, + "input": 128000 }, "cost": { "input": 0, @@ -19484,7 +19633,8 @@ "temperature": true, "limit": { "output": 32000, - "context": 128000 + "context": 144000, + "input": 128000 }, "cost": { "input": 0, @@ -19500,7 +19650,8 @@ "temperature": true, "limit": { "output": 64000, - "context": 128000 + "context": 144000, + "input": 128000 }, "cost": { "input": 0, @@ -23696,22 +23847,6 @@ "cache_write": 0 } }, - "openai/gpt-5.3-chat": { - "reasoning": false, - "tool_call": true, - "attachment": true, - "temperature": false, - "limit": { - "output": 16384, - "context": 128000 - }, - "cost": { - "input": 1.75, - "output": 14, - "cache_read": 0, - "cache_write": 0 - } - }, "morph/morph-v3-large": { "reasoning": false, "tool_call": false, @@ -23792,22 +23927,6 @@ "cache_write": 0.38 } }, - "minimax/minimax-m2.7-highspeed": { - "reasoning": true, - "tool_call": true, - "attachment": true, - "temperature": true, - "limit": { - "output": 131100, - "context": 204800 - }, - "cost": { - "input": 0.6, - "output": 2.4, - "cache_read": 0.06, - "cache_write": 0.375 - } - }, "recraft/recraft-v2": { "reasoning": false, "tool_call": false, @@ -24517,23 +24636,6 @@ "cache_write": 0 } }, - "qwen/qwen3.5-plus": { - "reasoning": true, - "tool_call": true, - "attachment": false, - "temperature": true, - "limit": { - "output": 65536, - "context": 1000000 - }, - "cost": { - "input": 0.4, - "output": 2.4, - "cache_read": 0, - "cache_write": 0, - "reasoning": 2.4 - } - }, "claude-3-5-haiku-latest": { "reasoning": false, "tool_call": true, @@ -25778,23 +25880,6 @@ "cache_write": 0 } }, - "google/gemini-3-pro-image-preview": { - "reasoning": true, - "tool_call": false, - "attachment": true, - "temperature": true, - "limit": { - "output": 32768, - "context": 65536 - }, - "cost": { - "input": 2, - "output": 12, - "cache_read": 0, - "cache_write": 0, - "reasoning": 12 - } - }, "z-ai/glm-4-32b": { "reasoning": false, "tool_call": true,