Z.ai: GLM-5-Turbo
Chatz-ai/glm-5-turboGLM-5-Turbo is a foundation model deeply optimized for the OpenClaw scenario. It has been specifically optimized for the core requirements of OpenClaw tasks since the training phase, enhancing key capabilities such as tool invocation, command following, timed and persistent tasks, and long-chain execution.
200K context window
128K max output tokens
Released: 2026-03-16
Supported Protocols:openaianthropic
Available Providers:Zhipu
Capabilities:Function CallingReasoningPrompt CachingWeb Search
Pricing
| Type | Price |
|---|---|
| Input Tokens | $1.2/M |
| Output Tokens | $4/M |
| Cache Read | $0.24/M |
| Web Search | $0.005/R |
Code Examples
from openai import OpenAIclient = OpenAI(base_url="https://api.ofox.ai/v1",api_key="YOUR_OFOX_API_KEY",)response = client.chat.completions.create(model="z-ai/glm-5-turbo",messages=[{"role": "user", "content": "Hello!"}],)print(response.choices[0].message.content)
Related Models
Frequently Asked Questions
Z.ai: GLM-5-Turbo on Ofox.ai costs $1.2/M per million input tokens and $4/M per million output tokens. Pay-as-you-go, no monthly fees.
Z.ai: GLM-5-Turbo supports a context window of 200K tokens with max output of 128K tokens, allowing you to process large documents and maintain long conversations.
Simply set your base URL to https://api.ofox.ai/v1 and use your Ofox API key. The API is OpenAI-compatible โ just change the base URL and API key in your existing code.
Z.ai: GLM-5-Turbo supports the following capabilities: Function Calling, Reasoning, Prompt Caching, Web Search. Access all features through the Ofox.ai unified API.