By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Why AI Is Reshaping Computer System Design, And Just About Everything Else
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Why AI Is Reshaping Computer System Design, And Just About Everything Else
News

Why AI Is Reshaping Computer System Design, And Just About Everything Else

News Room
Last updated: 2026/02/05 at 8:42 AM
News Room Published 5 February 2026
Share
Why AI Is Reshaping Computer System Design, And Just About Everything Else
SHARE

For the past few years, AI has dominated any discussion of technology trends. Sure, PCs are faster and longer-lasting, phones can snap eye-popping photos, and TVs look better than ever. But these changes have been mostly incremental. The biggest change is the emergence of large language models, chatbots, and, though still early, AI agents.

This has had a big impact on system design. The latest PCs and mobile phones now use neural processors (NPUs) to run AI models locally.  But the change in servers and the data centers that run them has been even more dramatic. While we’re likely to continue to need traditional CPU-based servers for a very long time, all of the new applications require a large number of graphics processing units (GPUs), even if they aren’t doing graphics.

Of course, this has resulted in a variety of new and enhanced chips to run these applications.  Nvidia, which started the trend toward more general-purpose use of GPUs and on whose chips most of the important AI training still takes place, has released new chips designed for data centers every year. Longtime rival AMD has joined the competition, particularly in the past year, with credible entries in its Instinct line. And it seems like all the big cloud computing vendors—the hyperscalers—have dedicated and proprietary chips to run their own AI applications, with most taking a focus on inferencing, including Google’s TPUs, Amazon’s Trainium, and Microsoft’s Maia. All of these have introduced new versions in the past few months. 

This has left traditional CPU architectures behind, forcing a redesign of AI applications, systems, and data centers. Indeed, the anticipated demand for new data centers populated with such chips is so great that there’s good reason to think they will require their own power sources.  We’ve even had the hyperscalers talking about building or restarting nuclear power plants.

One thing that has stood out to me recently is how this has impacted the design of systems for large data centers. Rather than the big cloud providers or a few big computer companies designing the systems, now we have the GPU makers—Nvidia first, now joined by AMD—not only designing new chips and boards, but actually full systems that contain all the components you need to build a rack of servers.

This all was very clear at CES 2026, where what caught my eye most wasn’t consumer electronics, the field that originally gave the show its name, but large-scale servers meant for huge data centers, including AMD’s Helios, powered by its newest Instinct graphics cards and Nvidia’s Vera Rubin, which has its next-generation architecture. What stood out to me was just how much both companies are now focusing on a systems approach, promising not only improved performance but also much greater efficiency. This may be the most consequential change we’ve seen in computer design in years.

Nvidia’s Vera Rubin

Nvidia CEO Jensen Huang (Photo by Patrick T. Fallon / AFP via Getty Images)

At CES, Nvidia CEO Jensen Huang focused on its Vera Rubin “AI Factory,” the next generation of its data center server platform. This is a complete systems design including not only the Rubin GPU and Vera ARM-based CPU, but also four other chips: the NVLink 6 Switch for scale-up networking (i.e., networking within the rack), Spectrum-X Ethernet photonics for scale-out networking (i.e., networking among racks), CX9 network interface cards, and BlueField 4 data processing units. A full Vera Rubin NVL72 has 72 GPUs, which Nvidia says delivers 50 petaflops at FP4 (four-bit precision inference), as well as 36 CPUs. Huang says the system is now in production.

Nvidia Vera Rubin

Nvidia Vera Rubin (Credit: Michael J. Miller)

“The amount of computation necessary for AI is skyrocketing,” Huang explained, noting that the top models are increasing the number of parameters by 10 times each year and test-time scaling or “reasoning” is using five times the number of tokens per year. Nvidia says its emphasis on codesigning all six chips in the system has allowed the company to deliver a 10x reduction in inference cost per token, compared with the previous generation of servers based on the current Blackwell GPU. In a later Q&A session, Huang said it would allow five times the throughput per watt.

Lisa Su, chair and chief executive officer of Advanced Micro Devices Inc. (AMD), holds an AMD EPYC

Lisa Su holds an AMD EPYC ‘Venice’ CPU (Credit: Bridget Bennett/Bloomberg via Getty Images)

AMD CEO Lisa Su was just as forceful in her CES keynote. She noted that AI has moved from using 1 zettaflop (sextillion floating point operations) in 2022 to more than 100 zettaflops in 2025. “We don’t have nearly enough compute for everything that we can possibly do.” 


Newsletter Icon

Newsletter Icon

Get Our Best Stories!

Your Daily Dose of Our Top Tech News


What's New Now Newsletter Image

Sign up for our What’s New Now newsletter to receive the latest news, best new products, and expert advice from the editors of PCMag.

Sign up for our What’s New Now newsletter to receive the latest news, best new products, and expert advice from the editors of PCMag.

By clicking Sign Me Up, you confirm you are 16+ and agree to our Terms of Use and Privacy
Policy.

Thanks for signing up!

Your subscription has been confirmed. Keep an eye on your inbox!

“To enable AI everywhere, we need to increase the world’s compute capacity, another 100 times over the next few years to more than 10 yottaflops (septillion floating point operations) over the next five years,” she said, noting that that would be 10,000 times more compute than we had in 2022, and a larger jump than we’ve ever seen in the history of computing.

amd helios

AMD Helios (Credit: Michael J. Miller)

AMD’s solution for this is a new version of its Instinct GPU, called the MI455X, and its own systems level product called Helios. It includes the Instinct MI455 accelerator with ultra fast HBM4 memory, along with a new Epyc CPU, codenamed Venice, as well as two Pensando DPUs (data processing units)—an existing Salian DPU plus an new one called Vulcano, which delivers faster UltraEthernet for scale-out computing. Again, this allows 72 GPUs per rack, which Su said will produce 2.9 Exaflops of AI compute, with 4,600 CPU cores, 18,000 GPU compute units, and 31 TB of HBM4 memory. This is due out later this year. Next year’s chip – the MI500 series – will be 10 times faster yet, and Su promised the combination would deliver a 1,000 times increase in AI performance over four years.

Why Systems Focus Is Important

A 1,000x improvement is pretty amazing—far faster than what we would have expected with Moore’s Law scaling, which in any case has actually been slowing. That’s due to both improvements on the chip level, including faster connections to memory; and on the systems level where all the different parts of the system are designed to work together.  And of course, the designers of AI software are also focused on making their systems more efficient.

All this improvement is crucial, because if AI and agent usage continues to grow, we’re going to need new systems and new data centers to run these applications. And you would expect that lower costs would only increase demand. But there’s only so many data centers that can be built quickly, so making them more efficient—getting more transactions in each server and using less power per transaction—is critical.

While all the big AI players and hyperscalers have promised new data centers, these are all limited by the amount of power that is available. New power generation is not growing nearly as fast as data center demand, so it’s important to see the emphasis on improving power efficiency.

Hybrid AI

Another big trend I noticed at CES was a widening focus on “hybrid AI.” AMD and Nvidia, as well as Intel, Lenovo, and Qualcomm all used this phrase at their keynotes and press events. But exactly what that means varies.

lenovo rack

Lenovo rack (Credit: Michael J. Miller)

Nvidia talked about three kinds of AI: training, inference, and simulation.  AMD, Intel, and Lenovo all talked about both data center servers and servers on “the edge” providing inference.  AMD, Intel, and Qualcomm all talked about running local AI models on PCs, something Microsoft has also been talking about in recent months. And Qualcomm and Lenovo (with its Motorola line) talked about running AI on phones.

The focus on AI everywhere was a clear indication of just how much AI has come to dominate the tech discussion. It seemed like every consumer electronics product I saw at the show has some kind of AI angle, though most were just new versions of things that had been out for years (such as improved image processing on televisions) or just another name for software. 

But there’s no question that devices, the applications we run, traditional personal software, and business applications are all undergoing some major shifts, driving by the AI emphasis.  A lot has changed already, but the bigger changes are yet to come.

About Our Expert

Michael J. Miller

Michael J. Miller

Former Editor in Chief


Experience

Michael J. Miller is chief information officer at Ziff Brothers Investments, a private investment firm. From 1991 to 2005, Miller was editor-in-chief of PC Magazine,responsible for the editorial direction, quality, and presentation of the world’s largest computer publication. No investment advice is offered in this column. All duties are disclaimed. Miller works separately for a private investment firm which may at any time invest in companies whose products are discussed, and no disclosure of securities transactions will be made.

Until late 2006, Miller was the Chief Content Officer for Ziff Davis Media, responsible for overseeing the editorial positions of Ziff Davis’s magazines, websites, and events. As Editorial Director for Ziff Davis Publishing since 1997, Miller took an active role in helping to identify new editorial needs in the marketplace and in shaping the editorial positioning of every Ziff Davis title. Under Miller’s supervision, PC Magazine grew to have the largest readership of any technology publication in the world. PC Magazine evolved from its successful PCMagNet service on CompuServe to become one of the earliest and most successful web sites.

As an accomplished journalist, well versed in product testing and evaluating and writing about software issues, and as an experienced public speaker, Miller has become a leading commentator on the computer industry. He has participated as a speaker and panelist in industry conferences, has appeared on numerous business television and radio programs discussing technology issues, and is frequently quoted in major newspapers. His areas of special expertise include the Internet and its applications, desktop productivity tools, and the use of PCs in business applications. Prior to joining PC Magazine, Miller was editor-in-chief of InfoWorld, which he joined as executive editor in 1985. At InfoWorld, he was responsible for development of the magazine’s comparative reviews and oversaw the establishment of the InfoWorld Test Center. Previously, he was the west coast bureau chief for Popular Computing, and senior editor for Building Design & Construction. Miller earned a BS in computer science from Rensselaer Polytechnic Institute in Troy, New York and an MS in journalism from the Medill School of Journalism at Northwestern University in Evanston, Illinois. He has received several awards for his writing and editing, including being named to Medill’s Alumni Hall of Achievement

Latest By Michael J. Miller

Read Full Bio

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article What is Lockdown Mode on iPhone and how do you enable it? What is Lockdown Mode on iPhone and how do you enable it?
Next Article Here’s how Apple may protect the display of the iPhone Fold – 9to5Mac Here’s how Apple may protect the display of the iPhone Fold – 9to5Mac
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Get an Amazon 4-Series 4K Fire TV in any size for up to 0 off
Get an Amazon 4-Series 4K Fire TV in any size for up to $180 off
News
Debian’s CI Data No Longer Publicly Browseable Due To LLM Scrapers / Bot Traffic
Debian’s CI Data No Longer Publicly Browseable Due To LLM Scrapers / Bot Traffic
Computing
Best wireless earbuds in 2026 reviewed and rated
Best wireless earbuds in 2026 reviewed and rated
Gadget
Big Changes Are Coming To Apple In 2026 – BGR
Big Changes Are Coming To Apple In 2026 – BGR
News

You Might also Like

Get an Amazon 4-Series 4K Fire TV in any size for up to 0 off
News

Get an Amazon 4-Series 4K Fire TV in any size for up to $180 off

3 Min Read
Big Changes Are Coming To Apple In 2026 – BGR
News

Big Changes Are Coming To Apple In 2026 – BGR

13 Min Read
Higgsfield launches ‘vibe’ editor for creating motion graphics –  News
News

Higgsfield launches ‘vibe’ editor for creating motion graphics – News

4 Min Read
Comer: Bill Gates subpoena ‘highly likely’ if he refuses cooperation in Epstein probe
News

Comer: Bill Gates subpoena ‘highly likely’ if he refuses cooperation in Epstein probe

0 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?