MiniMax: MiniMax M2.5
Chatminimax/minimax-m2.5MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1 to extend into general office work, reaching fluency in generating and operating Word, Excel, and Powerpoint files, context switching between diverse software environments, and working across different agent and human teams. Scoring 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, and 76.3% on BrowseComp, M2.5 is also more token efficient than previous generations, having been trained to optimize its actions and output through planning.
200K context window
131K max output tokens
Released: 2026-02-12
Supported Protocols:openaianthropic
Available Providers:MiniMaxAliyun
Capabilities:Function CallingReasoningPrompt CachingWeb Search
Pricing
| Type | Price |
|---|---|
| Input Tokens | $0.3/M |
| Output Tokens | $1.2/M |
| Cache Read | $0.03/M |
| Cache Write | $0.375/M |
Code Examples
from openai import OpenAIclient = OpenAI(base_url="https://api.ofox.ai/v1",api_key="YOUR_OFOX_API_KEY",)response = client.chat.completions.create(model="minimax/minimax-m2.5",messages=[{"role": "user", "content": "Hello!"}],)print(response.choices[0].message.content)
Related Models
Frequently Asked Questions
MiniMax: MiniMax M2.5 on Ofox.ai costs $0.3/M per million input tokens and $1.2/M per million output tokens. Pay-as-you-go, no monthly fees.
MiniMax: MiniMax M2.5 supports a context window of 200K tokens with max output of 131K tokens, allowing you to process large documents and maintain long conversations.
Simply set your base URL to https://api.ofox.ai/v1 and use your Ofox API key. The API is OpenAI-compatible — just change the base URL and API key in your existing code.
MiniMax: MiniMax M2.5 supports the following capabilities: Function Calling, Reasoning, Prompt Caching, Web Search. Access all features through the Ofox.ai unified API.