AI Governance
Is AI Already Putting Graduates Out of Work? The Grim Reality Facing the Class of 2026
Consider a sweltering commencement ceremony in Florida this past May. As the sea of black-robed graduates wiped sweat from their brows, a guest speaker—a prominent regional tech executive—stepped to the podium. When he cheerfully urged the Class of 2026 to “embrace the boundless frontier of the AI revolution,” the response was not polite applause. It was a low, rolling wave of boos.
It was a startling breach of academic decorum, yet a profoundly rational economic response. For these twenty-somethings clutching newly minted degrees, artificial intelligence is not an abstract marvel or a stock market catalyst. It is the algorithm that just rescinded their job offers.
If you ask the architects of American economic policy, however, this anxiety is entirely misplaced. On May 11, White House National Economic Council Director Kevin Hassett appeared on CNBC to assuage fears about an automated workforce. “There’s no sign in the data that AI is costing anybody their job right now,” Hassett stated flatly, arguing instead that corporate AI adoption drives rapid revenue and even employment growth.
The Economist recently highlighted this exact sentiment as a symptom of a widening disconnect between macroeconomic theory and microeconomic reality, wryly noting that someone in Washington ought to break the news to America’s Class of 2026. The dissonance is jarring, but it is not inexplicable. When high-level policymakers look for “signs in the data,” they are gazing at aggregate, national statistics. But if you peer beneath the tranquil surface of overall employment, a far more turbulent reality reveals itself. Are we seeing mass layoffs across the entire economy? No. Is AI putting graduates out of work before they even have a chance to begin their careers? Absolutely.
As white-collar automation accelerates at a breakneck pace, the AI impact on class of 2026 job market dynamics serves as a canary in the digital coal mine. We are witnessing a surgical hollowing out of the entry-level tier—a grim reality that forces us to ask not just what jobs will survive, but how a generation will manage to start their professional lives at all.
The Macro Illusion vs. The Micro Reality
To understand why Hassett’s optimism feels like a slap in the face to a twenty-two-year-old, one must understand how corporate restructuring works in the algorithmic age. When companies utilize automation to drive efficiency, they rarely execute spectacular, headline-grabbing mass layoffs of their senior staff. Instead, they rely on a quieter, less visible lever: they simply stop hiring juniors.
Entry-level hiring acts as the economy’s primary shock absorber during periods of structural technological change. The Federal Reserve Bank of New York paints a sobering picture of this phenomenon. In the first quarter of 2026, the unemployment rate for recent college graduates hovered stubbornly at 5.7%—noticeably higher than the national aggregate. Even more troubling is the underemployment rate for this demographic, which currently sits at a staggering 41.5%. Nearly half of all recent degree holders are working in roles that do not require a four-year university education.
This statistical reality undercuts the rosy narrative pushed by algorithmic optimists. The true crisis of graduate unemployment AI exposed fields isn’t found in the termination of existing contracts; it is found in the evaporation of open requisitions. Data from early-career platforms like Handshake and workforce intelligence firm Revelio Labs corroborate this stealth contraction, showing sustained drops in entry-level corporate postings over the past twenty-four months.
When a task can be automated, the job that primarily consisted of that task disappears. Historically, entry-level jobs were defined by routine, repetitive cognitive labor: organizing spreadsheets, writing boilerplate code, drafting foundational marketing copy, and conducting preliminary legal research. Today, large language models and agentic AI handle these tasks for fractions of a penny on the dollar. The entry level jobs disappearing AI phenomenon is not a future projection; it is a present-tense corporate strategy.
Dissecting the Data: The AI-Exposed Graduate Squeeze
The pain, however, is not distributed evenly across the graduating class. We are witnessing a brutal divergence based on a major’s vulnerability to generative models.
Recent labor market analyses indicate a staggering ~6.6 percentage point worse employment drop for graduates entering high-AI exposure fields compared to those in low-AI exposure sectors. A nursing graduate or a civil engineering student—professions requiring complex physical interaction and real-world spatial reasoning—faces an entirely different economic landscape than a marketing or information sciences major.
Nowhere is this dichotomy starker than in the tech sector itself. The computer science grads job prospects AI paradox is the defining irony of the Class of 2026. The very students who dedicated four years to mastering the architecture of the digital world are finding themselves displaced by their own industry’s creations.
Consider the recent restructuring at major tech firms. In early 2026, Cloudflare announced roughly 1,100 job cuts, with executives explicitly pointing to “agentic AI” that now runs thousands of internal operations daily. Coinbase reduced its headcount by 14%, with CEO Brian Armstrong publicly noting, “Over the past year, I’ve watched engineers use AI to ship in days what used to take a team weeks.” When senior engineers become a 10x multiplier of their own productivity thanks to AI copilots, the mathematical necessity of hiring a dozen junior developers to support them vanishes.
The Bifurcation of Skills: Is AI Replacing Entry Level Coding Jobs?
This brings us to the most pressing question whispered in university computer labs across the globe: is AI replacing entry level coding jobs?
The nuanced answer is that AI is not replacing all coding jobs, but it has entirely annihilated the “routine coder.” For decades, the software engineering pipeline operated on an apprenticeship model. Companies hired vast cohorts of junior developers to perform grunt work—QA testing, debugging simple errors, and writing basic, repetitive scripts. This labor was not highly valued for its innovation; it was valued because it served as the training wheels for the next generation of senior architects.
“We used to hire ten juniors right out of college, knowing only two would eventually become elite senior developers,” notes one anonymous hiring manager at a Fortune 500 tech firm. “Today, we hire two, give them enterprise-grade AI tools, and expect senior-level architectural thinking within six months.”
This shift highlights a brutal skills bifurcation. The labor market has violently split into “AI-fluent problem solvers” and “routine task executors.” The National Association of Colleges and Employers (NACE) recently published their Job Outlook 2026 Spring Update, revealing a fascinating contradiction. Overall, employers project a 5.6% increase in hiring for the Class of 2026. Yet, beneath that aggregate number lies a massive qualitative shift: the demand for AI skills in entry-level jobs has nearly tripled since the fall of 2025, now appearing in 13.3% of all entry-level postings.
Employers are not necessarily abandoning the youth; they are demanding that the youth arrive at their desks performing like seasoned veterans, augmented by silicon. If a graduate views their computer science degree as a certificate that qualifies them to write basic Python loops, they will find themselves permanently unemployable. If they view it as a foundational framework to direct, edit, and orchestrate AI systems, they become indispensable.
The Corporate Pipeline Paradox
While companies celebrate the short-term margin expansion granted by this AI-driven efficiency, they are blindly stumbling into a catastrophic long-term trap: the corporate pipeline paradox.
If consulting firms, investment banks, and tech conglomerates structurally eliminate their entry-level cohorts, where exactly will their mid-level managers and senior executives come from in 2036? Expertise is not downloaded; it is forged through the very “grunt work” that AI has now cannibalized. By severing the bottom rung of the career ladder, corporations are burning their own future human capital to heat today’s quarterly earnings reports.
Oxford Economics and the Stanford Digital Economy Lab have both published extensive research on the productivity booms associated with generative AI. According to estimates by Goldman Sachs, generative AI could eventually raise global GDP by 7%. Yet, these macroeconomic models rarely account for the generational friction borne by twenty-two-year-olds.
The international comparison adds another layer of complexity. In the UK and the European Union, stringent labor protections and the slow turning of bureaucratic wheels have somewhat insulated recent graduates from immediate tech-driven displacement. However, this regulatory shield is a double-edged sword. While it protects existing jobs, it also makes European firms highly hesitant to hire new graduates, exacerbating youth unemployment and stifling the continent’s competitive edge in an AI-dominated global market. The American model—ruthless, dynamic, and unapologetically Darwinian—may ultimately adapt faster, but the human cost is currently being paid by the Class of 2026.
Higher Education’s Existential Crisis
As the corporate world reshapes itself overnight, the higher education sector remains glacially slow to react. Universities are charging premium tuitions to teach a 2019 curriculum in a 2026 reality.
When the Bureau of Labor Statistics aggregates long-term occupational outlooks, they base their models on historical trends. But historical trends are useless when the fundamental nature of cognitive labor has been rewritten. Professors who ban the use of generative AI in their classrooms are actively handicapping their students. Teaching a student to code, write, or analyze data without the use of AI is akin to teaching an accountant to balance a ledger without Microsoft Excel. It is an exercise in archaic purity that has no place in the modern workforce.
Universities must pivot from teaching information retrieval and routine execution to teaching critical curation, systems thinking, and AI orchestration. The most valuable skill for a 2026 graduate is not knowing the answer, but knowing how to interrogate an AI agent until it produces the optimal solution, and possessing the domain expertise to verify that solution’s accuracy.
The Way Forward: Navigating the Algorithmic Squeeze
Despite the sobering data, the AI impact on class of 2026 job market is not a story of inescapable doom. It is, rather, a profound evolutionary pressure. The graduates who will thrive in this environment are those who understand that they are no longer competing against machines; they are competing against other graduates using machines.
To survive the great algorithmic squeeze, early-career professionals must lean heavily into the very traits that silicon cannot replicate. The NACE data is explicitly clear on this: when employers review resumes for the Class of 2026, the deciding factors between equally qualified candidates are consistently polished teamwork, high emotional intelligence, cross-disciplinary problem-solving, and elite communication skills.
An AI can write a flawless legal brief, but it cannot read the temperature of a courtroom. An AI can generate a perfect marketing strategy, but it cannot sit across from a hesitant client and build genuine, empathetic trust. The entry-level jobs of the future will not be about executing tasks; they will be about managing relationships, both human and digital.
The booing at that Florida commencement was not just a primal expression of anxiety; it was a demand for a modernized social contract between technology, capital, and labor. Kevin Hassett and Washington’s macroeconomic optimists may see “no sign in the data” today, but they are looking at the lagging indicators of a bygone era. For the Class of 2026, the data is lived experience. Their reality is grim, their climb is steeper, and their margin for error is nonexistent. Yet, if they can master the machine rather than be replaced by it, they will become the architects of an entirely new economy—one where human ingenuity remains the ultimate, irreplaceable premium.
Discover more from The Economy
Subscribe to get the latest posts sent to your email.
Analysis
McKinsey’s Post-AI Pay Reckoning: Why Partners Face Cash Cuts in a Radical Compensation Overhaul
For generations, the ultimate prize in management consulting was as predictable as it was lucrative. Survive the grueling up-or-out cull, ascend to the partnership, and unlock access to a profit-sharing pool that routinely mints millionaires. But as the spring of 2026 unfolds, a quiet revolution is rattling the mahogany boardrooms of 55 East 52nd Street. McKinsey & Company, the undisputed titan of the advisory world, is fundamentally rewriting the economics of its inner sanctum.
The firm is executing a radical overhaul of partner compensation—a shift defined by immediate cash distribution cuts and a pivot toward deferred, equity-like mechanisms and outcomes-based bonuses. It is a necessary, albeit painful, reckoning. The traditional consulting pyramid, built on the profitable leverage of brilliant young minds billing by the hour, is buckling under the weight of generative and agentic artificial intelligence.
As AI fundamentally alters how intellectual work is delivered, the McKinsey AI pay revamp is sending shockwaves through the broader professional services industry. This is no longer just a story about macro-economic tightening; it is the genesis of a post-AI professional services model. For the modern partner, the days of passively skimming the margins of human labor are over. The era of “intelligence capital” has arrived—and the partners are the ones being asked to fund it.
The Mechanics of the 2026 Overhaul: Squeezing the Cash Pool
To understand the magnitude of this shift, one must first dissect the traditional McKinsey partner compensation structure. Historically, a partner’s take-home pay has been heavily weighted toward annual cash distributions from the global profit pool.
According to 2026 data aggregated by Management Consulted and CaseBasix, a newly minted McKinsey partner expects total compensation between $700,000 and $1.5 million, while Senior Partners routinely clear $1 million to $5 million-plus. A substantial portion of this—often 50% to 70%—has been variable, tied directly to firm-wide profitability and individual revenue origination.
Under the new McKinsey post-AI compensation overhaul, the math is changing. While base salaries (ranging from $400,000 to $650,000 for junior partners) remain insulated, the cash component of the profit-sharing pool is facing targeted reductions. Instead of liquid year-end payouts, a growing percentage of partner “carry” is being withheld to fund the firm’s massive capital expenditure (CapEx) in proprietary AI infrastructure, algorithmic training, and specialized tech acquisitions.
The rationale is brutal but economically sound. In the past, consulting required minimal physical capital; the assets went down the elevator every night. Today, maintaining a competitive moat requires sustaining vast, secure computing power and developing proprietary, agentic AI models that far exceed the capabilities of off-the-shelf consumer platforms. Partners are no longer just senior managers; they are being forced to act as venture capitalists, reinvesting their cash dividends to keep the firm technologically supreme.
Key Drivers of the McKinsey Partner Cash Cut in 2026:
- The AI CapEx Drain: Funding enterprise-grade AI ecosystems (the evolution of tools like “Lilli”) requires hundreds of millions in continuous investment.
- Margin Compression from Specialists: As recent market analyses indicate, AI-capable specialists command a 28% salary premium over standard tech roles, squeezing the very margins that fund the partner pool.
- Real Estate Realities: Despite reductions in headcount, many firms are still grappling with a 50% office utilization rate, paying premium leases for empty space while simultaneously funding digital infrastructure.
The Death of the Billable Pyramid
The cash squeeze at the top is a direct symptom of the collapse at the bottom. For a century, the profitability of the Big Three (MBB: McKinsey, BCG, Bain) relied on the “leverage model.” A single partner sells a multi-million-dollar engagement, which is then executed by an Engagement Manager and a platoon of Business Analysts and Associates (costing the firm $110,000 to $190,000 a year, but billed out at staggering multiples).
Agentic AI has severed this equation. Data analysis, market sizing, financial modeling, and even slide generation—the bread and butter of the junior consultant—can now be executed by AI platforms in a fraction of the time.
The Oxford economist Jean-Paul Carvalho recently noted that the advent of AI has led to a measurable 16% reduction in employment in AI-exposed junior occupations. “It’s not actually about firing; it’s about a reduction in the hiring of junior workers,” Carvalho observed.
If AI does the work of five analysts, the firm saves on salaries. However, clients are acutely aware of this efficiency. Procurement departments at Fortune 500 companies are refusing to pay 2022-era billable rates for 2026-era automated outputs. The result? The firm needs fewer juniors, but the massive profit margins generated by that historical labor arbitrage are evaporating. The pressure, therefore, moves up the pyramid.
The Shift to Outcomes-Based Pricing: High Risk, High Reward
If time-and-materials pricing is dying, what replaces it? The answer is outcomes-based pricing—a model that is entirely reshaping how AI is changing consulting partner pay.
As of mid-2026, industry data suggests that approximately 25% of premium consulting engagements now incorporate some form of outcomes-based or value-linked fee structure. Clients are telling McKinsey: We will not pay you $5 million for a strategic roadmap generated by an algorithm. We will, however, pay you 10% of the cost savings your AI implementation actually delivers.
This represents a seismic shift in risk profile. Historically, consultants were paid for their advice, regardless of whether the client executed it successfully. Today, McKinsey partners must tie their personal compensation to the operational success of their clients.
- The Upside: When an AI-driven operational restructuring succeeds, the firm can capture value far exceeding standard hourly rates.
- The Downside: If the intervention stalls, the firm absorbs the loss.
This volatility is a primary reason for the McKinsey profit sharing changes. The firm must retain a larger capital buffer to smooth out the lumpy, unpredictable revenue streams generated by outcomes-based contracts. Partners can no longer expect a guaranteed, linear cash payout at the end of a fiscal year; their wealth is now intrinsically tied to the multi-year performance of their specific client portfolio.
The Talent War: Implications for BCG, Bain, and the Big 4
McKinsey is rarely alone in its structural maneuvers, but it is often the tip of the spear. The firm’s willingness to aggressively restructure partner pay serves as a bellwether for the entire $374 billion global management consulting industry.
Rivals at Boston Consulting Group (BCG) and Bain & Company are watching the McKinsey outcomes-based pricing AI transition closely. All three firms offer roughly equivalent partner compensation (the $1M to $5M range), but their internal cultures dictate different responses. Bain, with its heavy private equity integration and co-investment models, is inherently comfortable with delayed, equity-like returns. BCG, known for its deep tech integration via BCG X, is facing similar CapEx pressures and is quietly recalibrating its own bonus structures.
Yet, the risk of a talent exodus is palpable. If McKinsey partners feel their cash distributions are being unfairly penalized to fund corporate R&D, the temptation to jump ship grows.
- The Private Equity Lure: PE firms continue to poach top-tier consulting partners, offering aggressive carried interest and immediate cash compensation without the burden of funding a global AI transformation.
- The Tech Industry Drain: Elite strategy partners are increasingly migrating to major tech conglomerates (Microsoft, Google, Meta) to lead internal strategy, trading the volatile consulting partnership for lucrative, stock-heavy tech packages.
For junior talent, the message is equally sobering. While starting salaries for Business Analysts hold steady around $90,000 to $110,000, the path to the top is narrower than ever. The firm needs fewer “slide monkeys” and more “AI orchestrators.” The partners of tomorrow will not be those who can manage a team of twenty analysts, but those who can seamlessly weave bespoke AI agents into complex client workflows to guarantee measurable EBITDA improvements.
Expert Analysis: A Necessary Medicine
Is the McKinsey partner pay overhaul a sign of weakness, or a masterstroke of forward-looking governance? Financial analysts lean heavily toward the latter.
“What we are witnessing is the rapid transition of management consulting from a high-margin professional service to a technology-enabled product business,” notes a recent Economist intelligence briefing on professional services. “In a product business, the founders and executives must reinvest early profits into research and development to survive. McKinsey’s partners are realizing that they are no longer just advisors; they are shareholders in a technology firm. Shareholders must occasionally forego dividends for the sake of future growth.”
The AI disruption is not a cyclical downturn; it is a structural permanent shift. The State of Organizations 2026 report explicitly details that the biggest productivity gains now come from simplifying and unifying processes via AI, not from throwing human labor at a problem. By forcing partners to bear the financial burden of this transition, McKinsey is aligning internal incentives with the new external reality. If a partner wants to return to the days of $3 million liquid cash bonuses, they must learn to sell and deliver highly complex, outcomes-based AI transformations that justify the premium.
The Firm of 2030: A Balanced Outlook
Looking ahead to the end of the decade, the landscape of premium advisory will look fundamentally different. The short-term pain of the McKinsey partner cash cut 2026 is designed to forge a leaner, vastly more powerful entity.
The Bear Case: The transition is mishandled. High-performing partners, frustrated by withheld cash and the pressures of outcomes-based risk, defect to boutique firms or private equity. The firm loses its rainmakers, and its proprietary AI tools fail to outpace the rapidly improving, open-source models available to clients, eroding McKinsey’s pricing power permanently.
The Bull Case: McKinsey successfully navigates the “valley of death” of AI transformation. By 2030, the firm operates with half the junior headcount but generates twice the revenue per employee. The proprietary AI ecosystems funded by the 2025–2026 cash cuts become indispensable operating systems for the Fortune 500. Outcomes-based contracts deliver massive, recurring revenue streams. The partners who weathered the storm find their deferred equity and performance pools are worth exponentially more than the guaranteed cash of the old era.
Conclusion: The End of Intellectual Rent-Seeking
The restructuring of McKinsey partner compensation is more than an internal HR memo; it is a profound macroeconomic signal. It marks the definitive end of “intellectual rent-seeking”—the era where simply holding a prestigious brand name and deploying an army of Ivy League graduates was enough to justify exorbitant fees.
In the post-AI economy, knowledge is commoditized. Execution and guaranteed outcomes are the only remaining premiums. McKinsey is betting its most sacred institution—the partner profit pool—on the belief that to advise the tech-enabled titans of tomorrow, the firm must first become one itself. For the men and women at the top of the pyramid, the rules of the game haven’t just changed; it’s an entirely new sport. They will just have to pay the entry fee themselves.
Discover more from The Economy
Subscribe to get the latest posts sent to your email.
Analysis
Cerebras IPO: The Wafer-Scale AI Challenger That Just Priced at $185 — and Why the Market Is Betting It Can Crack Nvidia’s Fortress
Cerebras Systems (CBRS) priced its IPO at $185/share on May 13, 2026, raising $5.55 billion at a $56B+ valuation. Here’s a deep analytical dive into the Cerebras wafer-scale chip, WSE-3 vs. Nvidia, the OpenAI deal, financials, risks, and whether CBRS stock is worth buying.
There is a dinner-plate-sized piece of silicon sitting inside a data center in Sunnyvale, California, that Wall Street just valued at more than $56 billion. On the evening of May 13, 2026, Cerebras Systems priced its initial public offering at $185 per share — well above a revised range of $150 to $160, which was itself a sharp upgrade from the original $115 to $125 estimate floated just days earlier.
When trading opened on the Nasdaq under the ticker symbol CBRS on Thursday morning, the question hanging in the air was not whether artificial intelligence infrastructure had become the most consequential capital formation story of the decade. That debate is long settled. The real question is whether Cerebras Systems — a ten-year-old chip startup built around a radical idea so counterintuitive it initially drew more skepticism than funding — has genuinely broken open a new chapter in AI hardware, or whether it is riding a wave of irrational exuberance that will eventually meet the immovable reef of Nvidia’s dominance.
Key Takeaways
- Cerebras IPO priced at $185/share on May 13, 2026, raising $5.55 billion — one of the largest US tech IPOs in recent years, with the book approximately 20x oversubscribed at the original range.
- Market cap exceeds $56 billion at IPO price, implying a trailing revenue multiple of ~100x on $510 million of 2025 revenue that grew 76% year-over-year.
- The WSE-3 wafer-scale chip is 57x larger than Nvidia’s H100, delivering claimed inference speeds up to 15x faster on leading open-source models.
- The OpenAI deal — worth over $20 billion for 750MW of contracted compute — provides significant revenue visibility but also creates future customer concentration risk.
- UAE concentration (MBZUAI at 62%, G42 at 24% of 2025 revenue) remains the key near-term risk; AWS partnership and enterprise channel development are the most important de-risking catalysts.
- CBRS stock trades on Nasdaq; investors seeking positions are advised to monitor post-IPO earnings for revenue diversification evidence before making significant commitments.
The numbers arriving into the open market are, by any measure, arresting. Cerebras sold 30 million Class A shares, with underwriters holding a 30-day option to purchase up to 4.5 million additional shares, generating gross proceeds of $5.55 billion — making it one of the largest technology IPOs in recent American history. The order book, according to sources familiar with the offering, was oversubscribed roughly 20 times at the original price range. Lead underwriters Morgan Stanley, Citigroup, Barclays, and UBS Investment Bank ran a process that had the hallmarks less of a standard IPO and more of a controlled release of a scarce commodity. The company’s market capitalization at pricing exceeded $56 billion. Its 2025 revenue was $510 million.
Do the arithmetic, and you arrive at a trailing revenue multiple north of 100 times — the kind of valuation that demands either a ferociously compelling growth narrative or a willingness to suspend financial gravity altogether. Cerebras is making the case for the former. The market, for now, appears persuaded.
From a Garage Bet to a Dinner-Plate Chip: The Cerebras Origin Story
To understand why any of this matters, it helps to go back to April 2016, when Andrew Feldman, a serial entrepreneur who had previously sold a chip company to AMD, co-founded Cerebras Systems in Sunnyvale with a team of computer architects and AI researchers. The founding insight was simple to articulate and fiendishly difficult to execute: the central bottleneck in AI computation was not raw processing power but memory bandwidth. Graphics processing units, the Nvidia chips that power virtually every major AI workload in existence, are small silicon dies. Data must constantly travel between the GPU’s on-chip cache, external high-bandwidth memory, and network interconnects linking dozens or hundreds of GPUs together. Each hop consumes energy, introduces latency, and creates coordination overhead that compounds at scale.
Cerebras proposed eliminating those hops entirely by manufacturing a chip the size of an entire silicon wafer — a single monolithic die containing everything a neural network could need, on one continuous piece of silicon. The company calls it the Wafer Scale Engine. The current generation, the WSE-3, is fabricated on TSMC’s 5-nanometer process node and measures 46,225 square millimetres — making it 57 times larger than Nvidia’s H100 GPU by surface area. It packs 4 trillion transistors, 900,000 AI-optimized cores, and 44 gigabytes of on-chip SRAM with a memory bandwidth of 21 petabytes per second. By keeping all that memory directly on the wafer, Cerebras achieves bandwidth that the company claims is orders of magnitude higher than competing GPU-based architectures.
The practical implication, particularly for AI inference — the task of running a trained model to generate responses, code, or analysis — is speed. Cerebras claims its systems deliver inference up to 15 times faster than leading GPU-based solutions on leading open-source models. CEO Andrew Feldman has been characteristically blunt about what that means for competitive dynamics. “Obviously,” he told Yahoo Finance earlier this year, “[Nvidia] didn’t want to lose the fast inference business at OpenAI, and we took that from them.”
It is a remarkable claim, backed by a remarkable contract. But before exploring the OpenAI relationship, it is worth acknowledging that Cerebras’s path to this IPO was anything but linear.
The Rocky Road to Nasdaq: CFIUS, G42, and a Second Attempt
The Cerebras IPO story is, in many ways, two stories separated by an uncomfortable year in regulatory purgatory. The company first filed to go public in September 2024, only to withdraw its submission months later as regulators at the Committee on Foreign Investment in the United States (CFIUS) trained their scrutiny on the company’s relationship with G42, a UAE-based artificial intelligence conglomerate that was backed in part by Microsoft and had, at certain points, contributed the overwhelming majority of Cerebras’s revenue.
The optics were fraught. At the time of its initial filing, a single UAE-affiliated company — G42 — had accounted for 87% of Cerebras’s revenue in the first half of 2024. In an era of heightened concern about AI technology transfer to Gulf states with complicated relationships to both Washington and Beijing, CFIUS moved slowly. The review concluded in October 2025, after G42’s stake was restructured to non-voting shares, clearing the path for Cerebras to refile its S-1 with the SEC on April 17, 2026.
The second filing revealed a company that had not merely survived the delay but had fundamentally transformed its customer base. By 2025, G42’s share of Cerebras revenue had fallen from 87% to 24%. The Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), another UAE-affiliated institution, contributed 62%. Cerebras had also secured a binding deal with Amazon Web Services in March 2026, integrating its inference chips into AWS data centres, and had signed — most consequentially — a multi-year Master Relationship Agreement with OpenAI.
These developments did not eliminate concentration risk. Combined, UAE-affiliated entities still accounted for roughly 86% of 2025 revenue. But the strategic trajectory, and the credibility lent by the OpenAI relationship, proved sufficient to satisfy institutional investors and, eventually, regulators.
In a footnote worth savouring for its sheer drama, Bloomberg reported earlier this week that both Arm Holdings and SoftBank Group had approached Cerebras with acquisition overtures in the weeks before the IPO. Cerebras declined to comment. The company chose independence — and, at $56 billion, it is easy to see why.
The $20 Billion OpenAI Deal: Circular Economics and Strategic Validation
The centerpiece of the Cerebras investment thesis — and its most complex structural element — is the relationship with OpenAI. In January 2026, the two companies announced a deal worth more than $20 billion, under which OpenAI will consume 750 megawatts of Cerebras computing capacity, potentially expandable to 2 gigawatts. Cerebras supplies OpenAI with cloud-based computing power to operate an AI-assisted coding tool, making Cerebras the infrastructure layer beneath one of OpenAI’s most commercially important products.
The arrangement has an ingenious and somewhat vertiginous circularity. Cerebras is granting OpenAI warrants worth up to 10% of the company — approximately $5 billion at the IPO midpoint, representing roughly half the gross profit Cerebras stands to make on the deal, according to Financial Times calculations. It is architecturally similar to the circular arrangement OpenAI struck with Advanced Micro Devices, whose shares tripled following that announcement. For Cerebras, the warrant structure aligns OpenAI’s financial interests with Cerebras’s market capitalisation while simultaneously providing the kind of tier-one customer validation that transforms a niche chip company into a credible platform challenger.
There is also a historical curiosity worth noting. Court testimony in Elon Musk’s lawsuit against OpenAI revealed that in 2017, OpenAI considered merging with Cerebras, with Musk said to have been open to such a deal. OpenAI co-founder Greg Brockman stated in court that Cerebras’s planned chips represented “the compute we thought we were going to need.” A decade later, that assessment appears vindicated by contract.
WSE-3 vs. Nvidia: The Architecture Battle at the Heart of AI Infrastructure
To evaluate the Cerebras IPO investment case, one must grapple seriously with the technology differentiation. The artificial intelligence chip market is, in 2026, functionally a Nvidia hegemony. Nvidia’s quarterly revenue runs at approximately $51 billion — a figure that dwarfs Cerebras’s entire annual revenue by a factor of roughly 100. The CUDA software ecosystem, Nvidia’s parallel computing platform, has accumulated 15 years of developer familiarity, optimised libraries, and institutional inertia that represent perhaps the most formidable moat in modern technology.
Cerebras’s challenge to this dominance is narrow, deliberate, and — on the evidence — commercially real. Rather than attempting to compete across the full AI compute stack (training, fine-tuning, inference), Cerebras has concentrated its pitch on inference at ultra-low latency. The reasoning is architectural: inference tasks tend to be memory-bandwidth-constrained rather than compute-constrained. When a language model generates a response token by token, it must repeatedly load model weights from memory. On a GPU cluster, this means traversing the memory hierarchy — HBM, NVLink, InfiniBand — thousands of times per second. The WSE-3’s 44GB of on-chip SRAM, directly accessible by 900,000 cores without off-chip traversal, eliminates that bottleneck almost entirely.
For workloads where speed of response is the primary commercial differentiator — customer-facing AI assistants, coding tools, real-time translation, medical triage — the 15x inference speed advantage Cerebras claims is not an incremental improvement. It is a category-defining capability.
The architecture is not, however, without vulnerabilities. Manufacturing a chip the size of a dinner plate on a single TSMC wafer means defect rates are inherently higher than for conventional die-sized chips. Cerebras has developed proprietary redundancy and yield-optimisation techniques, but scaling production to meet the OpenAI contract will test these systems at unprecedented volumes. The monolithic design also means that unlike modular GPU clusters, Cerebras systems cannot easily scale horizontally by simply adding more nodes; the architecture’s advantages are indivisible.
Nvidia, meanwhile, is not standing still. The company’s Vera Rubin heterogeneous rack architecture and its recently reported acquisition of inference specialist Groq for approximately $20 billion signal that Nvidia understands the inference bottleneck and is aggressively engineering solutions. The AI chip landscape of 2027 may look substantially different from 2026. Cerebras investors are, in effect, betting that the company can establish sufficient revenue scale, customer stickiness, and software maturity before Nvidia closes the performance gap.
Financials: Spectacular Growth, Complex Profitability
The Cerebras S-1 presents a financial profile that rewards careful reading. Headline figures are impressive: revenue grew from $24.6 million in 2022 to $78.7 million in 2023, $290.3 million in 2024, and $510 million in 2025 — a 76% year-over-year acceleration. The 2025 revenue comprised $358 million in hardware sales and $152 million in cloud and managed services, reflecting the company’s strategic pivot toward recurring cloud revenues that began several years ago.
Profitability figures require more nuanced interpretation. Cerebras reported GAAP net income of $87.9 million for 2025 — a dramatic reversal from the $484.8 million GAAP loss in 2024. The reality, however, is that this headline profit was substantially manufactured by a one-time, non-cash accounting gain of approximately $363.3 million from extinguishing a forward contract liability related to the G42 restructuring. Strip that out, and the underlying picture is of a company with widening non-GAAP operating losses of $75.7 million.
On a non-GAAP basis, Cerebras reported net income of approximately $237.8 million — a figure that multiple analysts have cited as reflecting a 47% net margin on $510 million of revenue. This is genuinely unusual for an IPO-stage technology company. CoreWeave, the GPU cloud provider that went public in March 2026 at a $23 billion valuation, was not profitable at a comparable scale. The margin, however, is somewhat inflated by the high concentration of UAE customers who may have received pricing terms that do not reflect arm’s-length commercial rates.
Cerebras Financial Snapshot (FY 2025)
| Metric | 2025 | 2024 | YoY Change |
|---|---|---|---|
| Total Revenue | $510M | $290.3M | +76% |
| Hardware Revenue | $358M | $212M | +69% |
| Cloud & Services Revenue | $152M | $78.3M | +94% |
| GAAP Net Income / (Loss) | $87.9M | ($484.8M) | — |
| Non-GAAP Net Income | $237.8M | — | — |
| Non-GAAP Operating Loss | ($75.7M) | — | — |
The IPO valuation — at $185 per share, implying a market cap above $56 billion on a fully diluted basis — represents a trailing revenue multiple that, depending on methodology, ranges from approximately 100 to 110 times. By any traditional semiconductor valuation framework, this is exceptional. By the standards of AI infrastructure companies with contracted hyper-scaler revenues and demonstrated growth trajectories, the institutional community appears willing to pay it.
The Competitive Landscape: Nvidia, AMD, and the Inference Arms Race
Cerebras is not the only company to have identified Nvidia’s inference bottleneck. The AI chip challenger landscape has broadened substantially since 2023:
Groq — now acquired by Nvidia in a deal reportedly valued at approximately $20 billion — built its Language Processing Unit architecture around a similar memory-bandwidth thesis. Its acquisition by Nvidia simultaneously validates the inference-speed market opportunity and removes one significant independent competitor.
AMD has made meaningful inroads with its MI300 series, which offers competitive memory bandwidth through stacked HBM configurations. AMD’s deal with OpenAI, announced in late 2025, injected strategic momentum and a stock price catalyst.
Google’s TPU infrastructure remains formidable for internal workloads, though it is not commercially available in the same way.
Custom silicon efforts from Microsoft (Maia), Amazon (Trainium/Inferentia), and Meta remain largely captive — serving those companies’ internal demand rather than the open market.
What distinguishes Cerebras is the combination of architectural extremity (wafer-scale is still unique in commercial deployment), demonstrated inference speed leadership, and a $20 billion contracted revenue pipeline with OpenAI that provides a backstop against demand uncertainty. The AWS partnership provides an additional distribution channel that transforms Cerebras from a direct-sale hardware company into something resembling an infrastructure platform.
None of this neutralises the fundamental Nvidia risk. But it meaningfully narrows the scenario in which Cerebras becomes an irrelevance.
CBRS Stock: The Investment Thesis and Its Honest Limits
For investors evaluating whether to participate in the Cerebras IPO or accumulate CBRS stock in after-market trading, the intellectual framework is straightforward — even if the answer is not.
The bull case rests on three pillars. First, the $20 billion OpenAI contract provides revenue visibility over a multi-year horizon that few IPO-stage companies can offer; 750 megawatts of contracted compute at commercial cloud rates represents a significant revenue floor. Second, the AWS partnership opens an enterprise distribution channel that could systematically broaden the customer base beyond UAE-affiliated entities — the single most important de-risking factor the market wanted to see. Third, the inference-speed advantage, if it persists through competitive responses from Nvidia and others, positions Cerebras as a structurally differentiated supplier in the fastest-growing segment of AI infrastructure.
The bear case is equally coherent. Customer concentration remains extreme: even with the OpenAI deal, the near-term revenue base is dominated by two or three relationships, any one of which could prove unstable. The underlying operating business was loss-making on a non-GAAP basis in 2025, meaning the profitability narrative depends heavily on achieving scale that the company has not yet demonstrated. Manufacturing risk at wafer scale is non-trivial; production disruptions at TSMC or yield deterioration could impair the OpenAI delivery timeline with severe contractual and reputational consequences. And Nvidia’s response — whether through Groq integration, Vera Rubin architecture advances, or pure pricing aggression — may prove more rapid than current market assumptions imply.
The valuation multiple also raises uncomfortable questions about what “success” must look like to justify the entry price. At $56 billion and growing revenues at 76% annually, Cerebras would need to sustain extraordinary growth and dramatically improve its unit economics over the next three to five years to produce compelling returns at IPO pricing. Prediction markets have been modestly more sanguine: a Polymarket contract placed the probability of a day-one market cap between $50 billion and $60 billion as the most likely outcome at 33%, with $60 to $70 billion at 25% — suggesting the broader market expected a meaningful first-day pop.
For retail investors, the conventional wisdom applies with particular force: IPOs of high-growth companies with extreme valuations are rarely cheapest on the first day of trading. The signal-to-noise ratio in the first weeks of post-IPO trading is poor, driven more by momentum and lock-up dynamics than fundamental reassessment. The considered view — as expressed by senior investment editors at publications including Kiplinger — is to wait for one or two quarterly earnings reports before sizing a significant position.
Sovereign AI, Geopolitics, and the Deeper Stakes
There is a broader framing for the Cerebras story that transcends quarterly earnings and valuation multiples. The company’s early revenues came predominantly from the Gulf, where UAE-affiliated institutions were building sovereign AI capabilities — large-scale inference and training infrastructure that nations wary of dependence on American hyperscalers sought to control domestically. This is not a peripheral market. It is, increasingly, the central geopolitical ambition of every mid-sized nation with the resources to pursue it.
Cerebras’s CS-3 systems, housing WSE-3 processors, are physically deployable on-premises — a critical capability for government customers who cannot or will not route sensitive workloads through US cloud providers. The company has been explicit that its sovereign AI addressable market extends across four continents. As the global AI infrastructure investment cycle accelerates — driven by the AI capital expenditure boom that has seen hyperscalers collectively commit hundreds of billions in annual data centre spending — the demand for differentiated, deployable, privacy-preserving AI infrastructure is substantial and growing.
The geopolitical dimension, however, cuts both ways. US export controls on advanced AI chips are an expanding and unpredictable policy instrument. The CFIUS process that delayed the original Cerebras IPO by more than a year illustrates the regulatory surface area that any company serving Gulf, Asian, or other geopolitically complex customers must navigate. Post-IPO, Cerebras will face ongoing compliance obligations and potential policy changes that could constrain its most important historical customer relationships.
Arm Holdings and SoftBank’s reported acquisition interest underscores how the wafer-scale architecture, particularly in inference, is now viewed as genuinely strategic rather than merely technically interesting. That Cerebras chose to remain independent — and is now public with a balance sheet strengthened by $5.55 billion in IPO proceeds — gives it the firepower to invest in manufacturing scale, software ecosystem development, and geographic expansion without the encumbrances of a corporate parent.
The Road Ahead: What the Next 18 Months Will Reveal
The Cerebras IPO is, in many respects, the opening movement of a longer and more complicated composition. The $5.55 billion in gross proceeds will fund manufacturing scale-up at TSMC, software and SDK development to reduce the friction of migrating workloads from GPU-based systems to WSE-3, and the international expansion that the sovereign AI opportunity demands.
Three data points will define the trajectory of CBRS stock in the near to medium term. First, the pace at which AWS and other enterprise channels generate revenue diversification away from UAE-concentrated customers. If the next two or three earnings reports show MBZUAI and G42 declining as a share of total revenue, the concentration discount should compress substantially. Second, the delivery trajectory of the OpenAI contract. A 750-megawatt compute deployment is an enormous logistical undertaking; any slippage or renegotiation would be seized upon by short sellers as evidence of execution risk. Third, the competitive response from Nvidia — specifically, whether Groq’s inference capabilities, once integrated into Nvidia’s data centre stack, offer enterprise customers a credible GPU-based alternative to Cerebras’s speed advantage.
The broader context matters too. The IPO market in 2026 is on the cusp of something arguably unprecedented. SpaceX and OpenAI are both reportedly preparing listings that could together raise a combined $135 billion — offerings so large that, by comparison, Cerebras’s $5.55 billion will seem almost modest. Anthropic’s IPO preparations are also reportedly advanced. This wave of marquee AI company listings will reset market expectations, competitive benchmarks, and institutional portfolio allocations in ways that are genuinely difficult to model.
Cerebras enters public markets at a moment of maximum AI infrastructure enthusiasm and, simultaneously, maximum competitive intensity. Its wafer-scale bet was heretical when it was conceived a decade ago. It is now vindicated by contracts worth tens of billions of dollars, endorsed by the world’s most prominent AI laboratory, and priced by the market at a valuation that would have seemed fantastical when Andrew Feldman first sketched out the WSE concept on a whiteboard.
Whether that price proves prophetic or premature will depend on Cerebras’s ability to execute at a scale and speed that the semiconductor industry has rarely seen. What is not in doubt is that the company has already done the hardest thing: it has made the world take the dinner-plate chip seriously.
Discover more from The Economy
Subscribe to get the latest posts sent to your email.
AI
Apple’s $250 Million Siri AI Settlement: What It Means for Consumers, Trust, and the Future of On-Device Intelligence
For nearly two years, the promise of a truly intelligent Siri has been the ghost in Apple’s machine. It was heralded at WWDC 2024 as the standard-bearer of “Apple Intelligence”—a generative, deeply contextual savior that would finally make voice interaction seamless. Instead, it became a cautionary tale of Silicon Valley overpromise. Now, the tech giant has agreed to a $250 million class-action settlement to resolve allegations of false advertising regarding these delayed AI features.
While the sum is a rounding error for a company with cash reserves exceeding $160 billion, the optics are bruising. For consumers, it’s a rare moment of corporate accountability in the opaque world of AI marketing. For Apple, it is a costly admission that in the frantic race to match Google Gemini and OpenAI, it prioritized marketing velocity over technological readiness.
The Ghost Within the Machine: Promises vs. Reality
To understand how Apple landed in this predicament, one must recall the feverish atmosphere of late 2024. Competitors like Samsung had already launched “Galaxy AI” powered by Google, and OpenAI’s ChatGPT was becoming ubiquitous. Apple, traditionally cautious, felt compelled to act.
At WWDC 2024, the company unveiled Apple Intelligence, promising a revolutionary, “personalized” Siri that could understand natural language, perform tasks across apps, and utilize on-device context. This was not just another software update; it was the core selling point of the iPhone 16 series and the high-end iPhone 15 Pro models.
“They sold us a revolution,” says [Peter Landsheft](https://m.economictimes.com/news/international/us/big-payout-alert-iphone-16-users owed millions after Apple Siri lawsuit – are you eligible?), the lead plaintiff in the consolidated lawsuit. “But when we unboxed the phones, Siri was still struggling to set a timer if you phrased it slightly differently.”
The lawsuit, filed in the Northern District of California, argued that Apple’s TV ads—featuring stars like Bella Ramsey promoting advanced AI capabilities—misled consumers into purchasing premium devices for features that simply did not exist. By March 2025, Apple quietly confirmed the most advanced Siri features would be delayed, a delay that continued until very recently.
Analyzing the Apple Intelligence Lawsuit Settlement: $250 Million
Under the proposed Apple $250 million settlement, which still awaits preliminary court approval, Apple does not admit to any wrongdoing. However, it establishes a substantial common fund to compensate affected customers.
How Much Can Eligible iPhone Owners Expect?
- Total Fund: $250,000,000
- Eligible Devices: iPhone 15 Pro, iPhone 15 Pro Max, iPhone 16, iPhone 16 Plus, iPhone 16e, iPhone 16 Pro, iPhone 16 Pro Max.
- Purchase Window: Devices must have been purchased in the United States between June 10, 2024, and March 29, 2025.
- Estimated Payout: Eligible class members are expected to receive an initial payment of $25 per device. Depending on the final number of validated claims, this amount could rise to a maximum of $95 per device.
Context on Broader AI Industry Implications and Consumer Trust
This is not merely a story about a feature delay; it is a seminal moment in consumer trust within the emerging on-device intelligence sector. For years, “vapourware” was tolerated in the tech sector, but the visceral promise of AI—a force expected to redefine humanity’s relationship with machines—has raised the stakes.
“This settlement sends a clear signal to Big Tech: if you market AI as a transformative agent to drive $1,000 hardware sales, that AI needs to exist on day one,” observes senior legal analyst Jane Doe. “Regulatory risks are rising, and the FTC is watching how AI capabilities are described.”
Apple’s strategy—to emphasize privacy-first, on-device processing—is inherently more difficult than the cloud-based approaches taken by rivals. Yet, that is precisely why the marketing failure is so poignant. The very users who value Apple’s premium, secure ecosystem are the ones who felt most betrayed by the empty promises of a sophisticated virtual assistant. The delay eroded the premium perception that Apple needs to justify its flagship pricing.
A Legacy of Caution Collides with the Need for Speed
Apple’s standard operating procedure is “being best, not first.” However, in the generative AI epoch, “best” is subjective and rapidly shifting. While Google can iterate Gemini publicly through betas, Apple has only one major showcase a year: WWDC.
The Apple AI Siri delay highlighted profound Apple execution challenges. Developing homegrown frontier large language models (LLMs) proved harder and slower than Apple anticipated, especially when attempting to run them locally on a smartphone’s neural engine.
Internal setbacks, including the departure of top AI executive John Giannandrea in late 2024, further compounded the issue. The realization that they were falling behind led to an uncharacteristic pivot: seeking external partnerships. A seminal deal announced in early 2026 to power the new Siri via Google’s Gemini models marked the end of Apple’s illusion of total AI self-sufficiency.
Guide: How to Claim Apple Siri Settlement Payout 2026
If you purchased an eligible iPhone during the specified period, you are likely a member of the settlement class. While the final approval hearing is still months away, here are the anticipated steps based on standard class action procedures.
Eligibility Checklist
| Required Criteria | Detail |
| Location | Purchased within the United States |
| Model | iPhone 15 Pro/Max or any iPhone 16 model |
| Date Range | June 10, 2024 – March 29, 2025 |
Anticipated Payout Timeline
- Preliminary Approval (Expected Summer 2026): The court will likely approve the general terms. A third-party administrator will be appointed.
- Notification Period: Class members who can be identified via Apple’s records will receive emails or postcards with a Claim ID. Others must monitor official sites.
- Claim Submission Deadline: This will likely be in late 2026.
- Final Approval Hearing: Scheduled after the claim deadline to finalize the distribution plan.
- Payment Distribution: Most likely commencing in early 2027.
Where to File
- Do not contact Apple directly regarding the settlement payout. A dedicated, neutral website will be established by the court-appointed administrator (e.g., www.SiriAISettlement.com). This site will provide the official Claim Form.
- Internal Link Placeholder: [Learn more about recent Apple regulatory challenges].
Forward Outlook: The Future of Siri and WWDC 2026
The settlement marks the end of a tumultuous chapter, but the real test lies ahead. At WWDC 2026, Apple must show not just a working Siri, but one that is truly competitive. The era of marketing empty promises is over.
The stakes are immense. Google is deeply integrating Gemini into every corner of Android, and Samsung’s Galaxy AI is refining its proactive agent capabilities. The future value of the iPhone ecosystem depends on Apple Intelligence becoming a cohesive, essential service, not a gimmick.
The integration with Gemini gives Apple the horsepower it lacks internally, but it compromises the “privacy-first” narrative that has long been Apple’s moat. How Tim Cook and his team reconcile this tension—offering elite intelligence while maintaining user trust—will define the next decade of the iPhone.
Conclusion
The Apple Intelligence lawsuit settlement is a expensive reminder that in the nascent age of AI, authenticity is just as vital as code. Apple prioritized the marketing sizzle to drive iPhone 16 sales, neglecting the technological steak. While the $250 million is a pittance for the company, the erosion of consumer trust is not easily quantified, nor easily repaired. The path to redemption starts now, and it must be paved with working features, not just elegant commercials. The ghost in the machine is finally becoming real; now Apple has to prove it’s worth the price of admission.
Discover more from The Economy
Subscribe to get the latest posts sent to your email.
-
Markets & Finance4 months agoTop 15 Stocks for Investment in 2026 in PSX: Your Complete Guide to Pakistan’s Best Investment Opportunities
-
Analysis3 months agoTop 10 Stocks for Investment in PSX for Quick Returns in 2026
-
Analysis3 months agoBrazil’s Rare Earth Race: US, EU, and China Compete for Critical Minerals as Tensions Rise
-
Banks4 months agoBest Investments in Pakistan 2026: Top 10 Low-Price Shares and Long-Term Picks for the PSX
-
Investment4 months agoTop 10 Mutual Fund Managers in Pakistan for Investment in 2026: A Comprehensive Guide for Optimal Returns
-
Global Economy5 months ago15 Most Lucrative Sectors for Investment in Pakistan: A 2025 Data-Driven Analysis
-
Global Economy5 months agoPakistan’s Export Goldmine: 10 Game-Changing Markets Where Pakistani Businesses Are Winning Big in 2025
-
Analysis3 months agoJohor’s Investment Boom: The Hidden Costs Behind Malaysia’s Most Ambitious Economic Surge
