Close Menu
  • Home
  • Market News
    • Crude Oil Prices
    • Brent vs WTI
    • Futures & Trading
    • OPEC Announcements
  • Company & Corporate
    • Mergers & Acquisitions
    • Earnings Reports
    • Executive Moves
    • ESG & Sustainability
  • Geopolitical & Global
    • Middle East
    • North America
    • Europe & Russia
    • Asia & China
    • Latin America
  • Supply & Disruption
    • Pipeline Disruptions
    • Refinery Outages
    • Weather Events (hurricanes, floods)
    • Labor Strikes & Protest Movements
  • Policy & Regulation
    • U.S. Energy Policy
    • EU Carbon Targets
    • Emissions Regulations
    • International Trade & Sanctions
  • Tech
    • Energy Transition
    • Hydrogen & LNG
    • Carbon Capture
    • Battery / Storage Tech
  • ESG
    • Climate Commitments
    • Greenwashing News
    • Net-Zero Tracking
    • Institutional Divestments
  • Financial
    • Interest Rates Impact on Oil
    • Inflation + Demand
    • Oil & Stock Correlation
    • Investor Sentiment

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

Rolls-Royce Appoints Former bp CSO Ivanka Mamic as New Chief Sustainability Officer

January 15, 2026

Garmin Fenix 8 Smartwatch Review 2026

January 15, 2026

Chapo Sees Total LNG Project Restart Within Weeks

January 15, 2026
Facebook X (Twitter) Instagram Threads
Oil Market Cap – Global Oil & Energy News, Data & Analysis
  • Home
  • Market News
    • Crude Oil Prices
    • Brent vs WTI
    • Futures & Trading
    • OPEC Announcements
  • Company & Corporate
    • Mergers & Acquisitions
    • Earnings Reports
    • Executive Moves
    • ESG & Sustainability
  • Geopolitical & Global
    • Middle East
    • North America
    • Europe & Russia
    • Asia & China
    • Latin America
  • Supply & Disruption
    • Pipeline Disruptions
    • Refinery Outages
    • Weather Events (hurricanes, floods)
    • Labor Strikes & Protest Movements
  • Policy & Regulation
    • U.S. Energy Policy
    • EU Carbon Targets
    • Emissions Regulations
    • International Trade & Sanctions
  • Tech
    • Energy Transition
    • Hydrogen & LNG
    • Carbon Capture
    • Battery / Storage Tech
  • ESG
    • Climate Commitments
    • Greenwashing News
    • Net-Zero Tracking
    • Institutional Divestments
  • Financial
    • Interest Rates Impact on Oil
    • Inflation + Demand
    • Oil & Stock Correlation
    • Investor Sentiment
Oil Market Cap – Global Oil & Energy News, Data & Analysis
Home » AI Has Been All About GPUs. That’s Changing Fast.
U.S. Energy Policy

AI Has Been All About GPUs. That’s Changing Fast.

omc_adminBy omc_adminJanuary 8, 2026No Comments6 Mins Read
Share
Facebook Twitter Pinterest Threads Bluesky Copy Link


For years, Nvidia’s rise has been synonymous with one idea: GPUs are the engine of artificial intelligence. They powered the training boom that turned large language models from academic curiosities into trillion-dollar ambitions. But Nvidia’s $20 billion deal with Groq is an admission that the next phase of AI won’t be won by GPUs alone.

Groq makes a very different type of AI chip called a Language Processing Unit, or LPU. To understand why Nvidia spent so much, and why it didn’t simply build this technology itself, you have to look at where AI workloads are heading. The industry is moving from training models to running them in the real world. That shift has a name: inference.

Inference is what happens after a model is trained, when it answers questions, generates images, or carries on conversations with users. It’s becoming the dominant task in AI computing, and could dwarf the training market in the future, according to estimates recently compiled by RBC Capital analysts.

A chart on the outlook for AI training and inference markets

A chart on the outlook for AI training and inference markets

Structure Research/RBC Capital Markets



This matters because inference has very different needs than training. Training is like building a brain: it requires massive amounts of raw computing power and flexibility. Inference is more like using that brain in real time. Speed, consistency, power efficiency, and cost per answer suddenly matter far more than brute force.

That’s where Groq comes in. Founded by former Google engineers, Groq built its business around inference-only chips. Its LPUs are designed less like general-purpose factories and more like precision assembly lines. Every operation is planned in advance, executed in a fixed order, and repeated perfectly each time. That rigidity is a weakness for training, but a strength for inference, where predictability translates into lower latency and less wasted energy.

Alistair Barr, global tech editor of Business Insider, smiles at the camera while wearing a blue and white striped shirt.

Every time Alistair publishes a story, you’ll get an alert straight to your inbox!

Stay connected to Alistair and get more of their work as it publishes.

By contrast, Nvidia’s Graphics Processing Units, or GPUs, are designed to be flexible. They rely on schedulers and large pools of external memory to juggle many kinds of workloads. That flexibility is why GPUs won the training market, but it also creates overhead that slows inference down. As AI products mature and stabilize, that tradeoff becomes harder to justify.

“The tectonic plates of the semiconductor industry just shifted again,” Tony Fadell, creator of the iPod and an investor in Groq, wrote on LinkedIn recently. “GPUs decisively won the first wave of AI data centers: training. But inference was always going to be the real volume game, and GPUs by design aren’t optimized for it.”

Fadell calls this new breed of AI chips “IPUs,” or Inference Processing Units.

An explosion of different chips

TD Cowen analysts noted this week that Nvidia’s embrace of not just an inference-specific chip, but a whole new architecture, shows how large and mature the inference market has become.

Earlier AI infrastructure investments were made based on training-first buying decisions. The adage used to be “today’s training chips are tomorrow’s inference engines,” which favored Nvidia’s GPUs, but that’s no longer the case, the analysts added.

Instead, there will be an explosion of different chips inside future AI data centers, according to Chris Lattner, an industry visionary who helped develop the software for Google’s TPU AI chips, which Groq founder Jonathan Ross co-designed.

This move beyond GPUs is being driven by two trends that have been reinforced by Nvidia’s Groq deal, Lattner told me this week.

“The first is that ‘AI’ is not a single workload — there are lots of different workloads for inference and training,” he said. “The second is that hardware specialization leads to huge efficiency gains.”

“Humble move”

In a 2024 story (that aged very well), Business Insider warned readers that inference could be a vulnerability for Nvidia as rivals looked to fill this strategic gap. Cerebras built massive AI chips optimized for speed, claiming memory bandwidth thousands of times higher than Nvidia’s flagship GPU offering at the time. Google’s TPUs are designed to efficiently run bespoke AI workloads at blazing speeds. Amazon developed its own inference chip, Inferentia. Startups like Positron AI argued they could beat or match Nvidia’s inference performance at a fraction of the cost.

So Nvidia’s deal with Groq can be seen as a preemptive move. Rather than letting inference specialists chip away at its dominance, Nvidia chose to embrace a fundamentally different architecture.

Fadell described the deal as a “humble move” by Nvidia CEO Jensen Huang. “Many companies miss inflection points like this due to ‘Not Invented Here- driven egos,” Fadell added. “Jensen doesn’t; he saw the threat and made it work to his advantage.”

The economics of inference

The economics are compelling. Inference is where AI products make money. It’s the phase that proves whether hundreds of billions of dollars spent on data centers will ever pay off. As AWS CEO Matt Garman put it in 2024, if inference doesn’t dominate, “all this investment in these big models isn’t really going to pay off.”

Importantly, Nvidia isn’t betting on a single winner. GPUs will still handle training and flexible workloads. Specialized chips like Groq’s will handle fast, real-time inference. Nvidia’s advantage lies in owning the connective tissue — the software, networking, and developer ecosystem that lets these components work together.

“AI datacenters are becoming hybrid environments where GPUs and custom ASICs operate side-by-side, each optimized for different workload types,” RBC analysts wrote in a recent note, referring to Application-Specific Integrated Circuits such as Groq’s LPUs.

Some competitors argue the deal proves GPUs are ill-suited for high-speed inference. Others see it as validation of a more fragmented future, where different chips serve different needs. Nvidia’s Huang appears firmly in the second camp. By licensing Groq’s technology and bringing its team inside the tent, Nvidia ensures it can offer customers both the shovels and the assembly lines of AI.

Indeed, Nvidia has developed an NVLink Fusion technology that lets other custom chips connect directly to its GPUs, reinforcing this mixed-hardware future, the RBC Capital analysts noted.

“GPUs are phenomenal accelerators,” Andrew Feldman, CEO of Cerebras, wrote recently. “They’ve gotten us far in AI. They’re just not the right machine for high-speed inference. And there are other architectures that are. And Nvidia has just spent $20B to corroborate this.”

Sign up for BI’s Tech Memo newsletter here. Reach out to me via email at abarr@businessinsider.com.



Source link

Share. Facebook Twitter Pinterest Bluesky Threads Tumblr Telegram Email
omc_admin
  • Website

Related Posts

Garmin Fenix 8 Smartwatch Review 2026

January 15, 2026

How ‘Heated Rivalry’ Took Over the Internet

January 15, 2026

These Are the Leaders Nvidia Has Gained and Lost This Last Year

January 15, 2026
Add A Comment
Leave A Reply Cancel Reply

Top Posts

Citigroup must face $1 billion lawsuit claiming it aided Mexican oil company fraud

July 1, 20077 Views

LPG sales grow 5.1% in FY25, 43.6 lakh new customers enrolled, ET EnergyWorld

May 16, 20255 Views

Trump’s 100 days, AI bubble, volatility: Market Takeaways

December 16, 20075 Views
Don't Miss

Canopy Launches $2 Billion Platform to Scale Circular Textiles as Wood Supply Risks Rise

By omc_adminJanuary 15, 2026

New analysis warns that rising competition for wood fibre, tightening regulation, and climate disruptions threaten…

eBay Targets Net Zero by 2045 and Expands SBTi Validated Climate Plan

January 15, 2026

BBVA, ALTERRA Plan $1.2 Billion Global Climate Co Investment Fund

January 15, 2026

Abu Dhabi Awards Sustainability Labels to Private Sector Leaders at ADSW 2026

January 15, 2026
Top Trending

Rolls-Royce Appoints Former bp CSO Ivanka Mamic as New Chief Sustainability Officer

By omc_adminJanuary 15, 2026

Spain’s climate scientists subjected to ‘alarming’ rise in hate speech, minister warns | Spain

By omc_adminJanuary 15, 2026

L’Oreal Backs 13 Climate, Nature and Circularity Solutions Startups

By omc_adminJanuary 15, 2026
Most Popular

The 5 Best 65-Inch TVs of 2025

July 3, 202510 Views

The Layoffs List of 2025: Meta, Microsoft, Block, and More

May 9, 202510 Views

‘Looksmaxxing’ on ChatGPT Rated Me a ‘Mid-Tier Becky.’ Be Careful.

June 3, 20257 Views
Our Picks

Chapo Sees Total LNG Project Restart Within Weeks

January 15, 2026

Chapo Sees Total LNG Project Restart Within Weeks

January 15, 2026

Analyst Explains Why Feb NatGas Contract Collapsed Wednesday

January 15, 2026

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Facebook X (Twitter) Instagram Pinterest
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2026 oilmarketcap. Designed by oilmarketcap.

Type above and press Enter to search. Press Esc to cancel.