OpenAI signs deal with AI chipmaker Cerebras in a major multi‑year agreement to secure large‑scale computing capacity aimed at improving system responsiveness and real‑time performance. Under this arrangement, Cerebras will deliver 750 megawatts of compute power starting this year and continuing through 2028, making it one of the largest publicly disclosed supply agreements in the artificial intelligence sector to date.
The arrangement reflects rising demand for specialized computing infrastructure as AI platforms scale and as users expect faster and more reliable outputs. Both companies said the agreement is designed to reduce latency and improve processing speed for customers using advanced AI systems.
Large Scale Compute Supply Aims To Improve Real Time AI Performance
OpenAI signs deal to expand compute capacity, helping accelerate responses for tasks that currently take longer to process. This includes inference workloads supporting interactive applications, enterprise platforms, and security‑focused use cases where speed and accuracy are critical. Cerebras leadership compared the move toward real‑time inference to the way broadband transformed the internet, enabling faster and more engaging experiences.
Cerebras has spent more than a decade developing chips designed specifically for artificial intelligence workloads. Unlike traditional GPU based systems, the company uses wafer scale processors intended to handle large models with high throughput and low response times. Cerebras claims this design allows its systems to deliver faster inference compared with conventional clustered processor setups.
Cerebras Growth Reflects Rising Demand For Specialized AI Infrastructure
Interest in Cerebras has grown rapidly with the adoption of large language models and generative AI tools. The launch of ChatGPT in 2022 highlighted the limitations of existing compute infrastructure and underscored the need for alternatives capable of handling large‑scale, time‑sensitive workloads. In this environment, OpenAI signs deal with Cerebras to address rising demand for low‑latency compute as AI expands into real‑time applications, making responsiveness a central priority for developers and businesses.
Cerebras filed for an initial public offering in 2024 but has postponed the process multiple times. In the meantime, the company continued raising private funding and expanding its commercial footprint. Recent reports suggest Cerebras is in talks to secure an additional one billion dollars at a valuation of about 22 billion. Against this backdrop, OpenAI signs deal with Cerebras, building on a longstanding relationship after previously considering an acquisition before ultimately pursuing a partnership.
From a cybersecurity and infrastructure perspective, the agreement highlights how compute strategy has become a key part of platform resilience. Large AI providers increasingly rely on multiple hardware partners to reduce dependence on a single architecture or supplier. This approach supports system availability, performance stability, and risk management as AI services become more deeply integrated into business operations and digital security workflows.
OpenAI signs deal as part of a portfolio strategy designed to align the right systems with the right workloads. The addition of Cerebras infrastructure introduces a dedicated inference layer optimized for low‑latency performance. This capability is especially important for automated analysis, monitoring, and real‑time decision support systems that rely on immediate model responses.
As AI adoption continues to expand across industries, the size and scope of compute agreements are growing quickly. The OpenAI and Cerebras deal signals how infrastructure providers and AI developers are aligning more closely to meet performance expectations while preparing for continued growth in security-sensitive applications.

-600x600.jpg)




-1024x683.jpg)

