Google: Gemini 3.1 Flash Lite Preview
Chatgoogle/gemini-3.1-flash-lite-previewGemini 3.1 Flash Lite Preview is Google's high-efficiency model optimized for high-volume use cases. It outperforms Gemini 2.5 Flash Lite on overall quality and approaches Gemini 2.5 Flash performance across key capabilities. Improvements span audio input/ASR, RAG snippet ranking, translation, data extraction, and code completion. Supports full thinking levels (minimal, low, medium, high) for fine-grained cost/performance trade-offs. Priced at half the cost of Gemini 3 Flash.
1M context window
64K max output tokens
Released: 2026-03-03
Supported Protocols:openaianthropicgemini
Available Providers:Vertex
Capabilities:VisionFunction CallingReasoningPrompt CachingWeb SearchAudio InputVideo InputPDF Input
Pricing
| Type | Price |
|---|---|
| Input Tokens | $0.25/M |
| Output Tokens | $1.5/M |
| Audio Input | $0.5/M |
| Cache Read | $0.025/M |
| Cache Write | $1/M |
| Cache Write (1 hour) | $1/M |
| Cached Audio | $0.05/M |
| Web Search | $0.014/R |
Code Examples
from google import genaiclient = genai.Client(api_key="YOUR_OFOX_API_KEY",http_options={"api_version": "v1beta", "url": "https://api.ofox.ai/gemini"},)response = client.models.generate_content(model="google/gemini-3.1-flash-lite-preview",contents="Hello!",)print(response.text)
Related Models
Frequently Asked Questions
Google: Gemini 3.1 Flash Lite Preview on Ofox.ai costs $0.25/M per million input tokens and $1.5/M per million output tokens. Pay-as-you-go, no monthly fees.
Google: Gemini 3.1 Flash Lite Preview supports a context window of 1M tokens with max output of 64K tokens, allowing you to process large documents and maintain long conversations.
Simply set your base URL to https://api.ofox.ai/v1 and use your Ofox API key. The API is OpenAI-compatible — just change the base URL and API key in your existing code.
Google: Gemini 3.1 Flash Lite Preview supports the following capabilities: Vision, Function Calling, Reasoning, Prompt Caching, Web Search, Audio Input, Video Input, PDF Input. Access all features through the Ofox.ai unified API.