o4‑mini is a smaller, cost‑efficient reasoning model optimized for fast, high‑throughput usage; it excels in math, coding, and visual tasks, outperforms its predecessor o3‑mini, and supports significantly higher usage limits
The company that provides the model
The number of tokens you can send in a prompt
The maximum number of tokens a model can generate in one request
The cost of prompt tokens sent to the model
The cost of output tokens generated by the model
When the model's knowledge ends
When the model was launched
Capability for the model to use external tools
Ability to process and analyze visual inputs, like images
Support for multiple languages
Whether the model supports fine-tuning on custom datasets
o4-mini has a cost structure of $1.10 per million input tokens and $4.40 per million output tokens.
The input token cost for o4-mini is $1.10 per million input tokens.
The output token cost for o4-mini is $4.40 per million output tokens.
o4-mini supports a context window of up to 200,000 tokens.
o4-mini can generate up to 100,000 tokens in a single output.
o4-mini was released on April 16, 2025.
The knowledge cut-off date for o4-mini is May 31, 2024.
Yes, o4-mini supports vision capabilities.
Yes, o4-mini supports tool calling (functions).
Yes, o4-mini supports multiple languages, allowing it to handle input and output in several languages.
No, o4-mini does not support fine-tuning.
You can find the official documentation for o4-mini here:
o4-mini Documentation
Collaborate with thousands of AI builders to discover, manage, and improve prompts—free to get started.