By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Nvidia to Upgrade AI Chatbot Performance With New ‘LPU’ Chip
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Nvidia to Upgrade AI Chatbot Performance With New ‘LPU’ Chip
News

Nvidia to Upgrade AI Chatbot Performance With New ‘LPU’ Chip

News Room
Last updated: 2026/03/16 at 9:35 PM
News Room Published 16 March 2026
Share
Nvidia to Upgrade AI Chatbot Performance With New ‘LPU’ Chip
SHARE

To improve chatbot performance, Nvidia plans to sell a new kind of processor, an LPU, optimized to run large language models (LLMs). 

The “Nvidia Groq 3 LPU” chip was among seven upcoming chips Nvidia touted at the company’s annual GTC event, where it pitched the AI industry on why Nvidia’s chips continue to lead.

The LPU, or Language Processing Unit, comes from Nvidia’s deal this past December to license technology from a California AI company called Groq (not to be confused with the AI chatbot Grok from xAI). Founded in 2016, Groq issued earlier LPU chips specifically designed for LLMs to offer faster speeds and energy efficiency. The aim: To create an alternative to Nvidia’s enterprise GPUs, which can be used for a wider range of AI workloads.  

Nvidia now wants to pair the newly revealed Groq 3 LPU with the rest of the company’s next-generation AI chips, dubbed the “Vera Rubin” platform, which includes the upcoming Rubin GPU and Vera CPU tech for data centers. 

(Credit: Michael Kan)

Groq’s LPU chips use even faster SRAM (static RAM), instead of HBM (high-bandwidth memory) typically found on Nvidia’s GPUs. But on the downside, Groq’s LPUs can only offer “hundreds of megabytes” in SRAM, whereas HBM memory can span over a hundred gigabytes or more per chip.

That’s why a single Groq 3 LPU only contains 500MB of SRAM, while Nvidia’s upcoming Rubin GPU will feature 288GB of HBM4 memory. To compensate for the lower memory capacity, Nvidia is preparing to sell large batches of LPUs to work alongside the rest of its data center chips, giving AI companies a way to squeeze out even more performance. 


Newsletter Icon

Newsletter Icon

Get Our Best Stories!

Your Daily Dose of Our Top Tech News


What's New Now Newsletter Image

Sign up for our What’s New Now newsletter to receive the latest news, best new products, and expert advice from the editors of PCMag.

Sign up for our What’s New Now newsletter to receive the latest news, best new products, and expert advice from the editors of PCMag.

By clicking Sign Me Up, you confirm you are 16+ and agree to our Terms of Use and Privacy
Policy.

Thanks for signing up!

Your subscription has been confirmed. Keep an eye on your inbox!

Nvidia noted “the LPX rack with 256 LPU processors features 128GB of on-chip SRAM and 640TB/s of scale-up bandwidth. Deployed with Vera Rubin NVL72 (server unit), Rubin GPUs and LPUs boost decode by jointly computing every layer of the AI model for every output Token.”

Nvidia slide

(Credit: Michael Kan)

A data center could thus harness both the LPUs and Nvidia’s GPUs, dividing AI workloads between them to increase efficiency. Nvidia’s CEO, Jensen Huang, said the combined approach excels at helping AI companies boost performance with longer prompts.

Nvidia slide

(Credit: Michael Kan)

Combined, the LPUs and Rubin GPUs also promise to deliver up to a 35x increase in throughput when running a large language model with 1 trillion parameters, according to Nvidia’s benchmarks.

Recommended by Our Editors

“We’re in production with the Groq chip,” Huang said, adding that it’ll likely ship in Q3. Nvidia has contracted Samsung to manufacture the LPU. One analyst already expects Nvidia to ship out 4 to 5 million LPUs through 2026 and 2027.

Nvidia chips

(Credit: Michael Kan)

The new LPU and Vera Rubin systems will likely cost tens of thousands of dollars per chip, putting them far out of reach of consumers. Instead, expect the biggest AI companies, including OpenAI, Anthropic, and Meta, to adopt these technologies, which could power your chatbot queries or image-generation requests in the near future. 

At GTC, Nvidia also talked up Vera Rubin, about which the company has gone into detail before, including at January’s CES, where the company revealed the Rubin chips were in “full production.” Nvidia plans on shipping the Vera Rubin-related chips, including the new LPU chip, in this year’s second half.

About Our Expert

Michael Kan

Michael Kan

Senior Reporter


Experience

I’ve been a journalist for over 15 years. I got my start as a schools and cities reporter in Kansas City and joined PCMag in 2017, where I cover satellite internet services, cybersecurity, PC hardware, and more. I’m currently based in San Francisco, but previously spent over five years in China, covering the country’s technology sector.

Since 2020, I’ve covered the launch and explosive growth of SpaceX’s Starlink satellite internet service, writing 600+ stories on availability and feature launches, but also the regulatory battles over the expansion of satellite constellations, fights with rival providers like AST SpaceMobile and Amazon, and the effort to expand into satellite-based mobile service. I’ve combed through FCC filings for the latest news and driven to remote corners of California to test Starlink’s cellular service.

I also cover cyber threats, from ransomware gangs to the emergence of AI-based malware. Earlier this year, the FTC forced Avast to pay consumers $16.5 million for secretly harvesting and selling their personal information to third-party clients, as revealed in my joint investigation with Motherboard.

I also cover the PC graphics card market. Pandemic-era shortages led me to camp out in front of a Best Buy to get an RTX 3000. I’m now following how President Trump’s tariffs will affect the industry. I’m always eager to learn more, so please jump in the comments with feedback and send me tips.

Read Full Bio

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Sora vs. Kling vs. Runway: The Ultimate Video AI Showdown of 2026 – Chat GPT AI Hub Sora vs. Kling vs. Runway: The Ultimate Video AI Showdown of 2026 – Chat GPT AI Hub
Next Article Tencent Raises .27 Billion in Offshore Yuan Bonds as AI Spending Slows · TechNode Tencent Raises $1.27 Billion in Offshore Yuan Bonds as AI Spending Slows · TechNode
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Don’t Miss the Hype: Where to Stream Every 2026 Oscar Winner and Nominee
Don’t Miss the Hype: Where to Stream Every 2026 Oscar Winner and Nominee
News
Building Reliable AI Systems with AI Observability | HackerNoon
Building Reliable AI Systems with AI Observability | HackerNoon
Computing
5 AI features coming to your next car
5 AI features coming to your next car
Software
India’s scattered workforce: the chatbot keeping families in touch during emergencies
India’s scattered workforce: the chatbot keeping families in touch during emergencies
News

You Might also Like

Don’t Miss the Hype: Where to Stream Every 2026 Oscar Winner and Nominee
News

Don’t Miss the Hype: Where to Stream Every 2026 Oscar Winner and Nominee

9 Min Read
India’s scattered workforce: the chatbot keeping families in touch during emergencies
News

India’s scattered workforce: the chatbot keeping families in touch during emergencies

10 Min Read
Best TV deal: Save 0.01 on Insignia 55-inch Class F50 TV
News

Best TV deal: Save $170.01 on Insignia 55-inch Class F50 TV

3 Min Read
Why Hybrid Cars Make That Weird Humming Noise In Reverse – BGR
News

Why Hybrid Cars Make That Weird Humming Noise In Reverse – BGR

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?