diff --git a/src/lib/forbidden-free-models.ts b/src/lib/forbidden-free-models.ts index 67039998c..f658d773b 100644 --- a/src/lib/forbidden-free-models.ts +++ b/src/lib/forbidden-free-models.ts @@ -18,7 +18,6 @@ const forbiddenFreeModelIds: ReadonlySet = new Set([ 'liquid/lfm-2.5-1.2b-thinking:free', 'meta-llama/llama-3.2-3b-instruct:free', 'meta-llama/llama-3.3-70b-instruct:free', - 'minimax/minimax-m2.5:free', // only usable through kilo-auto 'mistralai/mistral-small-3.1-24b-instruct:free', 'nousresearch/hermes-3-llama-3.1-405b:free', 'nvidia/nemotron-3-nano-30b-a3b:free', diff --git a/src/lib/providers/minimax.ts b/src/lib/providers/minimax.ts index 0a1a7e704..22e48266b 100644 --- a/src/lib/providers/minimax.ts +++ b/src/lib/providers/minimax.ts @@ -7,7 +7,7 @@ export const minimax_m25_free_model: KiloFreeModel = { 'MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1 to extend into general office work, reaching fluency in generating and operating Word, Excel, and Powerpoint files, context switching between diverse software environments, and working across different agent and human teams. Scoring 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, and 76.3% on BrowseComp, M2.5 is also more token efficient than previous generations, having been trained to optimize its actions and output through planning.', context_length: 204800, max_completion_tokens: 131072, - status: 'hidden', + status: 'public', flags: ['reasoning', 'prompt_cache'], gateway: 'openrouter', internal_id: 'minimax/minimax-m2.5', diff --git a/src/tests/openrouter-models-sorting.approved.json b/src/tests/openrouter-models-sorting.approved.json index 871a00506..b77c0ba23 100644 --- a/src/tests/openrouter-models-sorting.approved.json +++ b/src/tests/openrouter-models-sorting.approved.json @@ -533,6 +533,59 @@ "default_parameters": {}, "isFree": true }, + { + "id": "minimax/minimax-m2.5:free", + "canonical_slug": "minimax/minimax-m2.5:free", + "hugging_face_id": "", + "name": "MiniMax: MiniMax M2.5 (free)", + "created": 1756238927, + "description": "MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1 to extend into general office work, reaching fluency in generating and operating Word, Excel, and Powerpoint files, context switching between diverse software environments, and working across different agent and human teams. Scoring 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, and 76.3% on BrowseComp, M2.5 is also more token efficient than previous generations, having been trained to optimize its actions and output through planning.", + "context_length": 204800, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000000", + "completion": "0.0000000", + "request": "0", + "image": "0", + "web_search": "0", + "internal_reasoning": "0", + "input_cache_read": "0.00000000" + }, + "top_provider": { + "context_length": 204800, + "max_completion_tokens": 131072, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "max_tokens", + "temperature", + "tools", + "reasoning", + "include_reasoning" + ], + "default_parameters": {}, + "isFree": true, + "settings": { + "included_tools": [ + "search_and_replace" + ], + "excluded_tools": [ + "apply_diff", + "edit_file" + ] + } + }, { "id": "xiaomi/mimo-v2-omni:free", "canonical_slug": "xiaomi/mimo-v2-omni:free",