The Cerebras Team
News

Cerebras Targets $3.5B IPO as Demand for AI Compute Surges

4 minute read
Michelle Hawley avatar
By
SAVED
Cerebras is seeking up to $3.5 billion in a Nasdaq IPO, testing Wall Street’s appetite for AI compute beyond Nvidia.

Key Takeaways

  • Cerebras plans to sell 28 million shares at $115 to $125 each in a Nasdaq IPO.
  • The AI chipmaker could reach a valuation of about $26.6 billion at the top of the range.
  • The offering would give public investors another way to bet on AI infrastructure beyond Nvidia.
  • Cerebras is shifting from selling hardware to operating cloud-based AI compute services.

Cerebras Systems is preparing to test just how much appetite public markets still have for AI infrastructure.

The Sunnyvale, California-based AI chipmaker is seeking to raise up to $3.5 billion in an initial public offering on the Nasdaq, according to an updated prospectus filed Monday. Cerebras plans to offer 28 million shares priced between $115 and $125 apiece, giving the company a potential valuation of roughly $26.6 billion at the top end of the range.

That would mark a step up from the company’s $23 billion private valuation in February, when it raised money from investors, including Advanced Micro Devices. It would also place Cerebras among the most closely watched technology IPOs of the current AI cycle — not because it's another software company promising AI productivity gains, but because it sits much closer to the physical bottleneck of the AI boom: compute.

Table of Contents

A Public-Market Test for AI Infrastructure

After a long slowdown in technology IPOs, AI has reopened parts of the market for companies that can plausibly claim a seat in the infrastructure stack. Generative AI has created intense demand for chips, data centers, cloud capacity and power — the less glamorous machinery behind chatbots, copilots and AI agents.

Cerebras aims to be one of the rare public-market alternatives to Nvidia’s dominance in AI infrastructure. Nvidia’s graphics processing units remain the default engine for much of the AI industry, but Cerebras has built its pitch around a different architecture: wafer-scale processors designed specifically for large-scale AI workloads.

Unlike conventional processors cut from a silicon wafer, Cerebras’ Wafer-Scale Engine uses an entire wafer as one giant processor. The company says the design brings compute and memory closer together, which can reduce some of the bottlenecks that slow AI training and inference.

Cerebras Wafer-Scale Engine
Cerebras Wafer-Scale Engine

That pitch now has to survive public-market scrutiny.

Related Article: OpenAI’s IPO Faces Questions Before It Even Begins

The Deal at a Glance

IPO detailCerebras' plan
ExchangeNasdaq Global Select Market
Expected tickerCBRS
Shares offered28 million
Price range$115 to $125
Potential gross proceedsUp to $3.5 billion
Potential valuationUp to about $26.6 billion
Underwriter optionAdditional 4.2 million shares

If underwriters exercise their option to buy another 4.2 million shares, Cerebras could bring in an additional $525 million at the top of the range. Morgan Stanley, Citigroup, Barclays and UBS Investment Bank are listed as lead book-running managers for the offering.

Notably, co-founder and CEO Andrew Feldman is not selling shares in the IPO. After the offering, he is expected to hold 10.3 million shares, worth as much as $1.28 billion if the IPO prices at the high end.

From Hardware Sales to Cloud Compute

Cerebras’ IPO is also a story about a business model in transition.

The company previously sought to go public in September of 2024, then withdrew the paperwork as it shifted away from a more straightforward hardware sales model. Its current strategy leans more heavily into cloud services powered by its own processors — a move that makes Cerebras look less like a traditional semiconductor vendor and more like an AI infrastructure provider.

That transition matters because customers increasingly want access to compute capacity, not just ownership of hardware. AI labs and enterprises are trying to run larger models, serve more users and lower inference costs without waiting for the next wave of Nvidia GPU capacity.

Cerebras’ biggest validation point is its OpenAI relationship. In January, the company said it would provide up to 750 megawatts of AI computing power to OpenAI through 2028 in a deal valued at more than $20 billion.

That gives Cerebras a marquee customer at the center of the AI boom. It also creates a familiar IPO question: how much of the company’s growth depends on a small number of very large customers?

Revenue Growth Gives Cerebras a Stronger Story

Unlike many AI companies selling future potential, Cerebras is entering the IPO process with meaningful recent financial momentum.

The company reported fourth-quarter revenue of $510 million, up about 76% from the year-earlier period, and net income of $87.9 million for the quarter. That profitability snapshot could help Cerebras stand out in a market that has become more selective about high-growth companies burning cash.

The comparison point is CoreWeave, the AI cloud provider that rents access to Nvidia GPUs and raised $1.5 billion in an IPO last year. CoreWeave’s listing showed that investors were willing to back AI infrastructure companies even when profitability was still under pressure. Cerebras is now asking whether the same market will reward a company trying to reduce dependence on Nvidia hardware rather than monetize it.

Related Article: The Chips Cold War: How GPUs Became the World’s Most Valuable Political Resource

Why the IPO Matters

The Cerebras IPO will be a referendum on three big questions facing the AI market:

  • Can specialized AI processors win investor confidence against Nvidia’s ecosystem advantage?
  • Will public investors reward companies tied to AI infrastructure as strongly as private investors have?
  • Can Cerebras turn its cloud pivot into a durable, scalable business?

For now, the company has the components investors tend to like: revenue growth, a massive AI customer, a differentiated technology story and exposure to one of the biggest capital spending cycles in tech.

But the risks are just as clear. Nvidia remains the center of gravity in AI infrastructure. Cloud services are expensive to build and operate. Large customer concentration can make revenue growth look more predictable than it really is. And AI valuations have already priced in a lot of future demand.

Learning Opportunities

Cerebras hopes public investors still want another way into the AI compute boom. The IPO roadshow will show whether Wall Street sees it as a true Nvidia alternative or just another expensive bet on demand that still has to materialize.

About the Author
Michelle Hawley

Michelle Hawley is an experienced journalist who specializes in reporting on the impact of technology on society. As editorial director at Simpler Media Group, she oversees the day-to-day operations of VKTR, covering the world of enterprise AI and managing a network of contributing writers. She's also the host of CMSWire's CMO Circle and co-host of CMSWire's CX Decoded. With an MFA in creative writing and background in both news and marketing, she offers unique insights on the topics of tech disruption, corporate responsibility, changing AI legislation and more. She currently resides in Pennsylvania with her husband and two dogs. Connect with Michelle Hawley:

Main image: Cerebras
Featured Research