Cerebras and Core42 Set New Performance Records with OpenAI’s gpt-oss-120B, Boosting AI Innovation Worldwide

Cerebras and Core42 Redefine AI Performance with OpenAI’s gpt-oss-120B, Enabling Real-Time Reasoning at Unprecedented Speed

In a groundbreaking development for the artificial intelligence landscape, Cerebras Systems, the world’s fastest AI provider, has partnered with Core42, a G42 company specializing in sovereign cloud and AI infrastructure, to deliver unparalleled performance for OpenAI’s gpt-oss-120B. This collaboration marks a new era of real-time reasoning and agentic workloads, empowering enterprises, researchers, and governments worldwide to harness cutting-edge AI capabilities at record-breaking speeds.

Through Core42’s AI Cloud and Compass API, enterprises can now access Cerebras Inference at an astonishing 3,000 tokens per second. This milestone not only sets a new benchmark for speed but also ensures ultra-low latency and significantly lower costs per token compared to traditional GPU-based systems. The partnership underscores both companies’ commitment to advancing open-source AI models while delivering scalable, enterprise-grade solutions that meet the demands of modern applications.

A Leap Forward for Open-Weight Models

OpenAI’s gpt-oss-120B is one of the most advanced open-weight models available today, offering exceptional reasoning power, long-context understanding (up to 128K tokens), and real-time capabilities. These features make it ideal for a wide range of use cases, from semantic search and code execution to automation and decision intelligence. By integrating this model into Core42’s AI Cloud, businesses gain seamless access to next-generation AI tools capable of driving innovation across industries.

Trevor Cai, Head of Infrastructure at OpenAI, emphasized the significance of this collaboration: “Together with Cerebras and Core42, we’re making our best and most usable open model available at unprecedented speed and scale. This will enable enterprises, researchers, and governments around the world to build real-time reasoning applications with extraordinary efficiency.”

Cerebras’ Wafer-Scale Technology Powers Breakthrough Performance

At the heart of this achievement lies Cerebras’ revolutionary wafer-scale engine (WSE) technology, which powers the CS-3 system. Unlike traditional GPU architectures, the WSE delivers deterministic, ultra-low-latency performance with memory optimization, ensuring consistent results even under heavy workloads. This architecture enables real-time inference for some of the largest AI models in existence, all while maintaining radically lower costs per token.

Andrew Feldman, CEO and co-founder of Cerebras, highlighted the impact of this technological advancement: “The latest chapter in our ongoing strategic partnership with Core42 brings the world’s most capable open-weight models directly into the hands of enterprises, researchers, and governments globally. Core42’s AI Cloud and Compass API make it effortless to leverage our inference performance, paving the way for a new generation of agentic workloads at the fastest speeds.”

For organizations seeking to deploy mission-critical AI applications, the combination of Cerebras’ hardware and Core42’s infrastructure provides unmatched scalability and flexibility. Whether experimenting with small-scale projects or rolling out full-scale production deployments, users can seamlessly transition between stages without compromising performance or efficiency.

Industry-Leading Price-Performance Ratio

Cerebras’ purpose-built AI infrastructure stands out not only for its speed and accuracy but also for its cost-effectiveness. With pricing structured as follows, enterprises can achieve optimal value for their investments:

  • Throughput: 3,000 tokens per second
  • Input Cost: $0.25 per million tokens
  • Output Cost: $0.69 per million tokens

These rates position Cerebras as the price-performance leader in the AI space, offering the lowest cost per token for OpenAI’s latest models. When combined with Core42’s AI Cloud, this translates to significant savings for businesses deploying large-scale AI solutions.

Unlocking New Possibilities for Enterprises

The integration of OpenAI’s gpt-oss-120B into Core42’s AI Cloud opens up endless possibilities for enterprise AI adoption. Organizations can now build powerful, reasoning-capable systems optimized for performance and cost-efficiency. Key benefits include:

  • Agentic AI at Scale: Develop intelligent, autonomous systems capable of handling complex tasks with precision and speed.
  • Enterprise-Scale Performance: Execute the most demanding workloads globally, enabling advanced automation and real-time experiences.
  • Industry-Leading Speed: Integrate gpt-oss-120B into applications such as knowledge retrieval, reasoning, and long-context generation with ease.

Kiril Evtimov, CEO of Core42 and Group CTO of G42, expressed enthusiasm about the launch: “By running OpenAI gpt-oss on Cerebras hardware within Core42’s AI Cloud and Compass API, we are setting a new benchmark for performance, flexibility, and compliance in AI. Our customers can now deliver innovative application capabilities by leveraging state-of-the-art open-weight models at the fastest speeds globally.”

Driving Global AI Innovation

This collaboration represents more than just a technical breakthrough—it embodies a shared vision for democratizing access to advanced AI technologies. By providing global enterprises and developers with instant access to high-performance AI models, Cerebras and Core42 are accelerating innovation across sectors. From healthcare and finance to education and government, organizations can now create transformative solutions that drive operational efficiency, enhance decision-making, and unlock new opportunities.

As the demand for real-time, reasoning-driven AI continues to grow, the partnership between Cerebras and Core42 sets a new standard for what is possible. With record-breaking performance, industry-leading economics, and a commitment to open innovation, they are poised to shape the future of AI for years to come.

In an age where speed, efficiency, and scalability define success, Cerebras and Core42 have delivered a solution that empowers businesses to stay ahead of the curve—ushering in a new era of AI-powered transformation.

About Core42

Core42, a G42 company, empowers individuals, enterprises, and nations to unlock the full potential of AI through its comprehensive enablement capabilities. As a leading provider of sovereign cloud, AI infrastructure, and services, our mission is to accelerate the achievements of others and help them reach their most ambitious goals.

To learn more, please visit www.core42.ai

About Cerebras Systems

Cerebras Systems is a team of pioneering computer architects, computer scientists, deep learning researchers, and engineers of all types. We have come together to accelerate generative AI by building from the ground up a new class of AI supercomputer. Our flagship product, the CS-3 system, is powered by the world’s largest and fastest commercially available AI processor, our Wafer-Scale Engine-3. CS-3s are quickly and easily clustered together to make the largest AI supercomputers in the world, and make placing models on the supercomputers dead simple by avoiding the complexity of distributed computing. Cerebras Inference delivers breakthrough inference speeds, empowering customers to create cutting-edge AI applications. Leading corporations, research institutions, and governments use Cerebras solutions for the development of pathbreaking proprietary models, and to train open-source models with millions of downloads. Cerebras solutions are available through the Cerebras Cloud and on-premises. For further information, visit cerebras.ai or follow us on LinkedIn, X and/or Threads.

Source link

Share your love