Cloud Computing

Groq and HUMAIN Launch OpenAI’s New Open Models Day Zero

Groq and HUMAIN Launch OpenAI's New Open Models Day Zero

Available worldwide with real-time performance, low cost, and local support in Saudi Arabia

Groq, the pioneer in fast inference, and HUMAIN, a PIF company and Saudi Arabia’s leading AI services provider, today announced the immediate availability of OpenAI’s two open models on GroqCloud. The launch delivers gpt-oss-120B and gpt-oss-20B with full 128K context, real-time responses, and integrated server-side tools live on Groq’s optimized inference platform from day zero.

Groq has long supported OpenAI’s open-source efforts, including large-scale deployment of Whisper. This launch builds on that foundation, bringing their newest models to production with global access and local support through HUMAIN.

“OpenAI is setting a new high performance standard in open source models,” said Jonathan Ross, CEO of Groq. “Groq was built to run models like this, fast and affordably, so developers everywhere can use them from day zero. Working with HUMAIN strengthens local access and support in the Kingdom of Saudi Arabia, empowering developers in the region to build smarter and faster.”

“Groq delivers the unmatched inference speed, scalability, and cost-efficiency we need to bring cutting-edge AI to the Kingdom,” said Tareq Amin, CEO at HUMAIN. “Together, we’re enabling a new wave of Saudi innovation—powered by the best open-source models and the infrastructure to scale them globally. We’re proud to support OpenAI’s leadership in open-source AI.”

Built for full model capabilities

To make the most of OpenAI’s new models, Groq delivers extended context and built-in tools like code execution and web search. Web search helps provide real-time relevant information, while code execution enables reasoning and complex workflows. Groq’s platform delivers these capabilities from day zero with a full 128k token context length.

Unmatched price-performance

Groq’s purpose-built stack delivers the lowest cost per token for OpenAI’s new models while maintaining speed and accuracy.

gpt-oss-120B is currently running at 500+ t/s and gpt-oss-20B is currently running at 1000+ t/s on GroqCloud.

Groq is offering OpenAI’s latest open models at the following pricing:

  • gpt-oss-120B: $0.15 / M input tokens and $0.75 / M output tokens
  • gpt-oss-20B: $0.10 / M input tokens and $0.50 / M output tokens

Note: For a limited time, tool calls used with OpenAI’s open models will not be charged. Learn more at groq.com/pricing.

Global from day zero

Groq’s global data center footprint across North America, Europe, and the Middle East ensures reliable, high-performance AI inference wherever developers operate. Through GroqCloud, OpenAI’s open models are now available worldwide with minimal latency.

Explore AITechPark for the latest advancements in AI, IOT, Cybersecurity, AITech News, and insightful updates from industry experts!

PR Newswire

PR Newswire empowers communicators to identify and engage with key influencers, craft and distribute meaningful stories, and measure the financial impact of their efforts. Cision is a leading global provider of earned media software and services to public relations and marketing communications professionals.

Related posts

AIS Renews Microsoft Partner Status, Boosts Cloud Expertise

Business Wire

UNIDO Announces AI Alliance with Huawei and Global Partners

Business Wire

BrainChip Launches Akida Cloud for Neuromorphic AI Access

Business Wire