What is the story about?
OpenAI has introduced two smaller versions of its flagship GPT-5.4 model GPT-5.4 mini and GPT-5.4 nano aimed at improving speed and cost efficiency for
high-volume AI tasks. The models are designed for developers and enterprises seeking scalable solutions for automation, coding, and data processing.
Focus on subagent architecture
The launch signals a shift in OpenAI’s strategy toward “subagent” systems, where a powerful model like GPT-5.4 handles complex reasoning, while smaller models execute repetitive or parallel subtasks.
This layered approach is intended to optimise performance while reducing operational costs, particularly for large-scale deployments.
“This architecture allows companies to balance intelligence with efficiency,” said an AI industry analyst. “Not every task requires a full-scale model.”
Performance and capabilities
According to OpenAI, GPT-5.4 mini delivers significant improvements over its predecessor in coding, reasoning, and multimodal understanding. Internal benchmarks indicate it approaches flagship-level performance in certain coding tasks while operating at more than twice the speed.
The GPT-5.4 nano model is positioned as the most lightweight and cost-effective option. It is designed for tasks such as data extraction, text classification, and customer support automation.
Both models support multimodal inputs, enabling them to process text and images, including interpreting complex user interface screenshots.
An OpenAI spokesperson said, “These models are built to deliver reliable performance at scale, particularly in production environments.”
Pricing and developer integration
OpenAI has priced the models to appeal to developers and enterprises managing large workloads. GPT-5.4 mini costs USD 0.75 per million input tokens and USD 4.50 per million output tokens, while GPT-5.4 nano is priced at USD 0.20 and USD 1.25 respectively.
The models integrate with OpenAI’s Codex environment, allowing developers to perform code analysis and file reviews at significantly lower cost compared to the flagship model.
Availability across platforms
GPT-5.4 mini is available through OpenAI’s API, Codex, and ChatGPT, where it can be accessed by free-tier users through specific features. It also functions as a fallback model for premium users when usage limits are reached.
GPT-5.4 nano, meanwhile, is currently limited to API access, targeting enterprise-scale applications and automated workflows.
Expanding AI model ecosystem
OpenAI has been expanding its portfolio of AI models to cater to different use cases, from high-performance reasoning systems to lightweight, cost-efficient tools.
The introduction of smaller models follows the broader industry trend of optimising AI for real-world deployment, where speed and cost are often critical constraints.
Impact on Industry
The release is expected to benefit businesses that rely on automation, including customer support, data processing, and software development.
By lowering costs and improving efficiency, the models could accelerate adoption of AI across industries, particularly among startups and enterprises operating at scale.













