Sponsored Links

Training vs. Inference



Training vs. Inference
By Jeff Brown, Editor, The Bleeding Edge

Perhaps it's the distraction of the holidays…
Or a few too many drinks at the holiday party.
There has been some widespread and nearly ubiquitous journalistic ineptitude with regard to some recent developments in artificial intelligence semiconductors.
The story, as it was presented, is that Meta (META) is in talks to team up with Alphabet (GOOGL).
The reporting says Meta intends to spend billions of dollars using Google's tensor processing units (TPUs) in Google's cloud-based services and ultimately procure the TPUs for Meta's own use.
One media outlet proclaimed that Google's pending deal with Meta is designed to "compete directly with NVIDIA in the AI chip business"… and that this move would "cast Google as a serious rival to semiconductor giant NVIDIA."
Another trotted out this beauty: "Has Google burst the NVIDIA bubble?"
Or how about, "Google may be looking to get in on NVIDIA's act."
All I could do was shake my head…
While "the markets" and Wall Street fell for it – hook, line, and sinker.
Which Chips for Which Application
Have a look at what has transpired over the last few days with Alphabet's and NVIDIA's (NVDA) share prices:
1-Month Charts of NVIDIA (NVDA) and Alphabet (GOOGL)
GOOGL spiked about 10% higher in the last few days, while NVDA tumbled as much as 4.6%. (Note: NVDA's share price is in BLUE and GOOGL's share price is in WHITE.)
If we were to just look at the chart above, we'd think the media was right.
Good news for GOOGL, bad news for NVIDIA, right?
The chart and the media show us how little they know about the basics of semiconductors, the applications they are designed for, and how the industry works.
We're exactly three years into the AI infrastructure boom… and they still haven't figured out the basics.
And it's not hard to figure out. It's written everywhere.
Take, for example, Google's Ironwood TPU, which represents its seventh-generation TPU.
In Google's own words,"Ironwood: Google Cloud's 7th-Generation TPU Engineered for Inference."
Google's Ironwood TPUs (GOLD squares) | Source: Google Cloud
The key word, of course, is inference.

Recommended Links


image

Trump's Hidden Fed Agenda

Trump's latest moves show he's preparing to reshape the Federal Reserve – and the value of the U.S. dollar. With key appointees already in place, the coming monetary reset could send gold soaring like it did in the 1970s, when it climbed 24X in under a decade. Click here to see how Trump's Fed shift could spark a new gold boom.


image

The $25 Trillion Tesla Story No One Is Telling You

While the media obsesses over Tesla’s slumping car sales… Jeff has uncovered a revolutionary AI breakthrough buried inside Tesla’s labs. One that could be 14X bigger than the AI boom that minted 600,000 new millionaires. Click here to read the $25 trillion Tesla story no one is telling you.


AI Training vs. AI Inference
Google's TPUs are primarily designed and optimized for inference – the running of AI applications.
Does the media even understand this basic distinction – beyond the copy/pasted definition?
Because this is a critically important distinction to make between AI inference and AI training – the training of massive foundational AI models, which entirely use GPUs – what I call the general-purpose workhorses of artificial intelligence.
Regular Bleeding Edge readers will know by now that NVIDIA owns about 90% of the GPU market for AI training, and Advanced Microdevices (AMD) owns the remaining 10%.
Just two companies control 100% of the AI training market for frontier AI models. That's it. They own it. Own shares in these two companies, and you own it all.
Google's TPUs do not compete with NVIDIA's or AMD's GPUs for training frontier AI models.
It's plain and simple, which is why all the reporting on this Google and Meta mega-deal has been so dead wrong. Even the implications of the deal are misunderstood.
Where Google's TPUs do complete is in the inference market.
It's TPUs compete against:
  • AMD's inference-specific GPUs
  • Amazon's Inferentia
  • Meta's own Meta Training and Inference Accelerator (MTIA)
  • Marvel's (MRVL) Inference Solutions
  • Microsoft's (MSFT) Azure Maia 100 Accelerator
  • Cerebras Wafer Scale Engine
  • Groq's language processing units (LPUs)
  • SambaNova's reconfigurable dataflow units (RDUs)
  • Tenstorrent Blackhole processors
  • d-Matrix Corsair
  • and a long list of other emerging smaller players.
The entire semiconductor industry – outside of NVIDIA and AMD – is focused on inference for two key reasons:
  • It would be nearly impossible to break into the duopoly on GPUs.

    NVIDIA and AMD have locked down 100% of the available manufacturing capacity of TSM for GPUs with order books that exceed a year.

    And every software developer is well-versed in NVIDIA's CUDA development environment or AMD's ROCm, which is AMD's open-source software for GPU computing.

  • Inference is a wide-open, greenfield market that is growing exponentially… where NVIDIA's GPUs are not power efficient for inference, and a duopoly doesn't exist.
Even Tesla (TSLA) is part of this race, although it doesn't make its own inference chips available to third parties presently.
Tesla historically has used NVIDIA chips to train its frontier models, just as xAI has done to train Grok.
However, for inference and video-specific AI training, Tesla developed its own application-specific semiconductors manufactured by either TSM or Samsung Electronics.
For Tesla, designing its own custom semiconductors for inference is a competitive advantage.
The Real Purpose of the Google/Meta Tie-Up
The real purpose of Google and Meta teaming up for more TPUs is simple…
Both need more purchasing power with Taiwan Semiconductor Manufacturing (TSM).
Increased purchasing power isn't just about better pricing. It's a negotiating position to gain a larger allocation to TSM's total manufacturing capacity.
Outside of the need for increased energy production to fuel AI factories, TSM is the single largest bottleneck in AI.
All the companies that I listed above pay TSM to manufacture their semiconductors. All of them.
And the key implication of Google and Meta partnering for more TPUs has nothing to do with being a competitive threat to NVIDIA.
It is telling us something entirely different.
Demand for inference semiconductors for artificial intelligence is skyrocketing.
And that means that the utilization of AI applications is experiencing exponential growth at a scale that is nearly impossible to understand.
Skyrocketing demand for inference is the proof that we're not just chasing a bubble to achieve AGI.
Both the consumer, enterprise, and public sector adoption of AI is torrid, material, and tied to concrete and measurable revenues and free cash flow.
Stick with us at The Bleeding Edge in 2026 and beyond, and ignore the media if you'd like to have an inside track on what is really happening in high tech. Not only will it save you a lot of time, but it will be both intellectually and financially profitable.
Happy Thanksgiving to all.
We have so much to be grateful for…
And we have so much to look forward to.
Jeff

Keep reading
It looks like some kind of ancient and powerful artifact… Seen, perhaps, in a science fiction epic – like...
It was inevitable that a bad actor would use this technology with malicious intent.
There's an inherent danger when companies like Google, Meta, and Anthropic train their AIs on falsities, political narratives, and...

Like what you're reading? Send your thoughts to feedback@brownstoneresearch.com.

Brownstone Research
1125 N Charles St, Baltimore, MD 21201
www.brownstoneresearch.com

To ensure our emails continue reaching your inbox, please add our email address to your address book.

This editorial email containing advertisements was sent to ahmedwithnour@gmail.com because you subscribed to this service. To stop receiving these emails, click here.

Brownstone Research welcomes your feedback and questions. But please note: The law prohibits us from giving personalized advice.

To contact Customer Service, call toll free Domestic/International: 1-888-493-3156, Mon-Fri, 9am-5pm ET, or email us here.

© 2025 Brownstone Research. All rights reserved. Any reproduction, copying, or redistribution of our content, in whole or in part, is prohibited without written permission from Brownstone Research.

0 التعليقات:

إرسال تعليق

Share With Friends

Bookmark and Share
 
recipes for healthy food © 2008 | Créditos: Templates Novo Blogger