Tuesday, 17 Jun 2025
  • My Feed
  • My Interests
  • My Saves
  • History
  • Blog
Subscribe
News Outlet
  • Home
  • Opinion

    Don’t panic! But there’s young women with phones in Parliament! And it creates a better politics

    By
    Holly Rankin

    Five years on: How Covid changed everything

    By
    Rubi Mote

    Presenter who started career in Oxfordshire to leave BBC after 28 years

    By
    Christeen Grumbles

    The politics behind Starmer’s decision to scrap NHS England

    By
    Elida Michaud

    Electioncast: Three battleground seats to watch

    By
    Crystal Andrews

    Rate-setter calls on Bank of England to adopt ‘activist’ approach to rates

    By
    Ali Lyon
  • Politics
    Mike Berners-Lee: Why dishonesty is destroying the planet | WTCTW Podcast

    Mike Berners-Lee: Why dishonesty is destroying the planet | WTCTW Podcast

    By
    Johnathon Menjivar
    Ex-Fox News star Steve Hilton launches run for governor in California with goal of taking on Kamala Harris

    Ex-Fox News star Steve Hilton launches run for governor in California with goal of taking on Kamala Harris

    By
    Lawanda Howe
    Joe Biden should stay out of politics for the good of his party, says former aide to Jill

    Joe Biden should stay out of politics for the good of his party, says former aide to Jill

    By
    Rhian Lubin
    Politics LIVE: Keir Starmer warns businesses of ‘economic impact’ from Trump tariffs

    Politics LIVE: Keir Starmer warns businesses of ‘economic impact’ from Trump tariffs

    By
    Clora Lupo
    Séance politics: Is it unpatriotic for Turnbull to question AUKUS?

    Séance politics: Is it unpatriotic for Turnbull to question AUKUS?

    By
    Bernard Keane
    What does the Le Pen verdict mean for the future of French politics?

    What does the Le Pen verdict mean for the future of French politics?

    By
    Diego Wrona
  • Health
    Knots Landing star tragically dies aged 63

    Knots Landing star tragically dies aged 63

    By
    Lyndia Redner
    Blake Lively and Ryan Reynolds’ lookalike family member steals the show at star-studded gala

    Blake Lively and Ryan Reynolds’ lookalike family member steals the show at star-studded gala

    By
    Nancie Pekar
    Demi Moore showcases incredibly slender frame after revealing she ‘tortured’ her body

    Demi Moore showcases incredibly slender frame after revealing she ‘tortured’ her body

    By
    Gaylene Motsinger
    Six Million Dollar Man Lee Majors turns 86

    Six Million Dollar Man Lee Majors turns 86

    By
    Rebecka Stoval
    Former NBA star revealed as the world’s highest paid athlete — despite retiring years ago

    Former NBA star revealed as the world’s highest paid athlete — despite retiring years ago

    By
    Tyisha Kazmierczak
    ‘Divine’ Tom Ford perfume that ‘receives so many compliments’ now £62 in flash sale

    ‘Divine’ Tom Ford perfume that ‘receives so many compliments’ now £62 in flash sale

    By
    Georgianna Drews
  • Business
    California Overtakes Japan to Become World’s Fourth-Largest Economy — But Challenges Loom

    California Overtakes Japan to Become World’s Fourth-Largest Economy — But Challenges Loom

    By
    Raleigh Paris
    Nissan Bleeds $5.26 Billion in Losses: Why The Japanese Automaker Is Struggling to Restructure

    Nissan Bleeds $5.26 Billion in Losses: Why The Japanese Automaker Is Struggling to Restructure

    By
    Alejandro Pekar
    Pope Francis’s Death Sparks Frenzy For Papal Thrillers — Why Millions Are Watching ‘Conclave’

    Pope Francis’s Death Sparks Frenzy For Papal Thrillers — Why Millions Are Watching ‘Conclave’

    By
    Tama Lupo
    Quick Facts About Leanna Lenee: Age, Career, Relationship with Travis Hunter, and More

    Quick Facts About Leanna Lenee: Age, Career, Relationship with Travis Hunter, and More

    By
    Bong Mongold
    ‘Final Wish’: Heartwarming Reason Why Pope Francis Chose Santa Maria Maggiore For His Burial

    ‘Final Wish’: Heartwarming Reason Why Pope Francis Chose Santa Maria Maggiore For His Burial

    By
    Tomi Kazmierczak
    China Will Teach Kids AI Across the Country

    China Will Teach Kids AI Across the Country

    By
    Laine Grumbles
  • 🔥
  • Business
  • Marketing
  • Investment
  • World
  • Health
  • Discover
  • Politics
  • Opinion
  • Innovation
Font ResizerAa
News OutletNews Outlet
  • My Saves
  • My Interests
  • My Feed
  • History
  • Travel
  • Opinion
  • Politics
  • Health
  • Technology
  • World
Search
  • Personalized
    • My Feed
    • My Saves
    • My Interests
    • History
  • Bookmarks
  • News
    • Opinion
    • Politics
    • Technology
    • Travel
    • Health
    • World
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
InnovationScienceTechnology

We heard you like HBM – Nvidia’s Blackwell Ultra GPUs have 288 GB of it

Tobias Mann
Last updated: March 18, 2025 6:35 pm
Tobias Mann
Share
We heard you like HBM – Nvidia’s Blackwell Ultra GPUs have 288 GB of it
SHARE

GTC Nvidia’s Blackwell GPU architecture is barely out of the cradle – and the graphics chip giant is already looking to extend its lead over rival AMD with an Ultra-themed refresh of the technology.

Announced on stage at Nvidia’s GPU Technology Conference (GTC) in San Jose, California, on Tuesday by CEO and leather jacket aficionado Jensen Huang, the Blackwell Ultra family of accelerators boasts up to 15 petaFLOPS of dense 4-bit floating-point performance and up to 288 GB of HBM3e memory per chip.

And if you’re primarily interested in deploying GPUs for AI inference, that’s a bigger deal than you might think. While training is generally limited by how much compute you can throw at the problem, inference is primarily a memory-bound workload. The more memory you have, the bigger the model you can serve.

According to Ian Buck, Nvidia veep of hyperscale and HPC, the Blackwell Ultra will enable reasoning models including DeepSeek-R1 to be served at 10x the throughput of the Hopper generation, meaning questions that previously may have taken more than a minute to be answered can now be done in as little as ten seconds.

With 288 GB of capacity across eight stacks of HBM3e memory onboard, a single Blackwell Ultra GPU can now run substantially larger models. At FP4, Meta’s Llama 405B could fit on a single GPU with plenty of vRAM left over for key-value caches.

To achieve this higher capacity, Nvidia’s Blackwell Ultra swapped last-gen’s eight-high HBM3e stacks for fatter 12-high modules, boosting capacity by 50 percent. However, we’re told that memory bandwidth remains the same at a still class-leading 8 TB/s.

If any of this sounds familiar, this isn’t the first time we’ve seen Nvidia employ this strategy. In fact, Nv is following a similar playbook to its H200, which was essentially just an H100 with faster, higher-capacity HBM3e onboard. However, this time around, with these latest Blackwells, Nvidia isn’t just strapping on more memory, it’s also juiced the peak floating-point performance by 50 percent – at least for FP4 anyway.

Nvidia tells us that FP8 and FP16/BF16 performance is unchanged from last gen.

More memory, more compute, more ‘GPUs’

While many have fixated on Nvidia’s $30,000 or $40,000 chips, it’s worth remembering that Hopper, Blackwell, and now its Ultra refresh aren’t one chip so much as a family of products ranging the gamut from PCIe add-in cards and servers to rack-scale systems and even entire supercomputing clusters.

In the datacenter, Nvidia will offer Blackwell Ultra in both its more traditional HGX servers and its rack-scale NVL72 offerings.

Nvidia’s HGX form factor has, at least for the past few generations, featured up to eight air-cooled GPUs stitched together by a high-speed NVLink switch fabric. However, this time it a new config it’s calling the B300 NVL16, which might sound like they’re cramming twice as many GPUs into a box, something they’re previously done with the HGX V100. In reality, Nvidia changed its mind and has decided to count the individual compute dies on the package as GPUs.

According to Nvidia, the Blackwell-based B300 NVL16 will deliver 7x the compute and 4x the memory capacity of its Hopper generation, which we’ve learned refers to its 80GB H100s and not the higher capacity H200 systems. By our calculation works out to 112 petaFLOPS of dense FP4 compute or about 7 petaFLOPS of dense FP4 per GPU die, or 14 petaFLOPS per SXM module. That’s quite a performance uplift, with each B300 die performing on par with the Blackwell B100-series chips announced last year.

Nvidia does appear to have done a fair bit of rounding with its memory claims. By our calculation the HGX B300 systems actually deliver closer to 3.6x more memory at 2.3 TBs versus 640GB of the HGX H100.

For even larger workloads, Nvidia will also offer the accelerators in its Superchip form-factor. Unlike last year’s GB200, the GB300 Superchip will pair four Blackwell Ultra GPUs with a combined 1,152GB of HBM3e memory with two 72-core Grace Arm-compatible CPUs.

Up to 18 of these Superchips can be stitched together using Nvidia’s NVLink switches to form an NVL72 rack-scale system. But rather than the 13.5 terabytes of HBM3e of last year’s model, the Grace-Blackwell GB300-based systems will offer up to 20 terabytes of vRAM. What’s more, Buck says the system has been redesigned for this generation with an eye toward improved energy efficiency and serviceability.

And if that’s still not big enough, eight of these racks can be combined to form a GB300 SuperPOD system containing 576 Blackwell Ultra GPUs and 288 Grace CPUs.

Where does this leave Blackwell?

Given its larger memory capacity, it’d be easy to look at Nvidia’s line-up and question whether Blackwell Ultra will end up cannibalizing shipments of the non-Ultra variant.

However, the two platforms are clearly aimed at different markets, with Nvidia presumably charging a premium for its Ultra SKUs.

In a press briefing ahead of Huang’s keynote address today, Nvidia’s Buck described three distinct AI scaling laws, including pre-training scaling, post-training scaling, and test-time scaling, each of which require compute resources to be applied in different ways. 

At least on paper, Blackwell Ultra’s higher memory capacity should make it well suited to the third one of these regimes, as they allow customers to either serve up larger models – AKA inference – faster or at higher volumes.

Meanwhile, for those building large clusters for compute-bound training workloads, we expect the standard Blackwell parts to continue to see strong demand. After all, there’s little sense in paying extra for memory you don’t necessarily need.

With that said, there’s no reason why you wouldn’t use a GB300 for training. Nvidia tells us the higher HBM capacity and faster 800G networking offered by its ConnectX-8 NICs will contribute to higher training performance. 

  • DeepSeek-R1-beating perf in a 32B package? El Reg digs its claws into Alibaba’s QwQ
  • AI bubble? What AI bubble? Datacenter investors all in despite whispers of a pop
  • AI running out of juice despite Microsoft’s hard squeezing
  • Nvidia won the AI training race, but inference is still anyone’s game

Competition

With Nvidia’s Blackwell Ultra processors expected to start trickling out sometime in the second half of 2025, this puts it in contention with AMD’s upcoming Instinct MI355X accelerators, which are in an awkward spot. We would say the same about Intel’s Gaudi3 but that was already true when it was announced.

Since launching its MI300-series GPUs in late 2023, AMD’s main point of differentiation was that its accelerators had more memory (192 GB and later 256 GB) than Nvidia’s (141 GB and later 192 GB), making them attractive to customers, such as Microsoft or Meta, deploying large multi-hundred- or even trillion-parameter-scale models.

MI355X will also see AMD juice memory capacities to 288 GB of HBM3e and bandwidth to 8 TB/s. What’s more, AMD claims the chips will close the gap considerably, promising floating-point performance roughly on par with Nvidia’s B200.

However, at a system level, Nvidia’s new HGX B300 NVL16 systems will offer the same amount of memory, and significantly higher FP4 floating-point performance. If that weren’t enough, AMD’s answer to Nvidia’s NVL72 is still another generation away with its forthcoming MI400 platform. 

This may explain why, during its last earnings call, AMD CEO Lisa Su revealed that her company planned to move up the release of its MI355X from late in the second half to the middle of the year. Team Red also has the potential to undercut its rival on pricing and availability, a strategy it’s used to great effect in its ongoing effort to steal share from Intel. ®

Updated at 15.55 UTC on March 19, 2025, to add

This article was updated to clarify the memory configuration of the HGX B300 NVL16. We had asked Nvidia about this following our earlier chat with them, and it transpires the silicon goliath changed the definition of a GPU and NVL naming convention for the B300 but confusingly not the GB300.

Never mind that, our friends over at The Next Platform have more here on Nvidia’s roadmap to 2028

Read More

Share This Article
Email Copy Link Print
Previous Article Euro techies call for sovereign fund to escape Uncle Sam’s digital death grip Euro techies call for sovereign fund to escape Uncle Sam’s digital death grip
Next Article How to Get Caught Blowing Up a Tesla
Leave a Comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Recipe Rating




Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
XFollow
InstagramFollow
LinkedInFollow
MediumFollow
QuoraFollow
- Advertisement -
Ad image

You Might Also Like

UK government told to get a grip on £23B tech spend
InnovationScienceTechnology

UK government told to get a grip on £23B tech spend

By
Lindsay Clark
Hisense QLED TVs are just LED TVs, lawsuit claims
InnovationScienceTechnology

Hisense QLED TVs are just LED TVs, lawsuit claims

By
Thomas Claburn
Euro techies call for sovereign fund to escape Uncle Sam’s digital death grip
InnovationScienceTechnology

Euro techies call for sovereign fund to escape Uncle Sam’s digital death grip

By
Dan Robinson
CVE fallout: The splintering of the standard vulnerability tracking system has begun
InnovationScienceTechnology

CVE fallout: The splintering of the standard vulnerability tracking system has begun

By
Jessica Lyons
News Outlet
Facebook Twitter Youtube Rss Medium

About US


News Outlet : Your instant connection to breaking stories and live updates. Stay informed with our real-time coverage across politics, tech, entertainment, and more. Your reliable source for 24/7 news.

Top Categories
  • World
  • Opinion
  • Politics
  • Tech
  • Health
  • Travel
Usefull Links
  • Advertise with US
  • Complaint
  • Privacy Policy
  • Cookie Policy
  • Submit a Tip

© 2025 News Outlet Network.  All Rights Reserved.

Welcome Back!

Sign in to your account

Username or Email Address
Password

Lost your password?