Samsung is one of the names of this February. They are expected to present the Galaxy S26, but they have something on the table that will be a shock not only to their coffers, but to the engine of the South Korean economy. We refer to high bandwidth memories because, in the midst of the RAM and SSD crisis, Samsung is prepared to mass manufacture HBM4 memories.
And it will be for AI, how could it be otherwise.
In short. The South Korean company has not confirmed it, but recent reports published by Reuters and local sources such as the Korea JoongAng Daily suggest that Samsung will begin mass manufacturing of HBM4 memory chips starting next week.
It will be the first of the three companies that dominate the production of memory chips (the others are the South Korean SK Hynix and the American Micron, which has left consumer RAM) to begin manufacturing these fundamental memories for artificial intelligence in large quantities.
Price of RAM: Samsung and SK Hynix have flatly refused ” width=”375″ height=”142″ src=”https://i.blogs.es/087496/memorias-ia-dram-stock/375_142.jpeg”/>
HBM4. This type of memory, as its name suggests, has enormous bandwidth. This is crucial for GPU needs, and while NVIDIA has stuck to GDDR memory for its graphics cards, AMD did flirt with stacked HBM chip technology for its Vega GPUs. However, it is not a technology for consumption, not because its performance is inadequate, but because it is too expensive.
Making HBM memory is more expensive than making traditional DRAM chips, but the advantages are there. With HBM4, for example, the density of the stacked chips allows for twice the bandwidth of the previous generation. This is key to transmitting more data per second, but they also consume up to 40% less energy than HBM3 memories.
NVIDIA. The most interested party is, as we have said on previous occasions, NVIDIA. And if NVIDIA benefits, practically the entire leading artificial intelligence industry will take advantage of it because its chips are what are currently moving the industry. It is estimated that Samsung memories will go to NVIDIA’s Vera Rubin acceleration systems
In fact, it has been reported that Jensen Huang himself has urged to accelerate and increase the production of these chips. Well, Huang has asked the entire semiconductor industry that manufactures components for his cards to step up, it is not something that only concerns Samsung.
Spearhead. According to a Korea KoongAng Daily source, “Samsung has the world’s largest production capacity and broadest product line. It has demonstrated a recovery in its technological competitiveness by becoming the first to mass produce the highest-performance HBM4 memory.”
Because, in this field, its main competitor, the neighboring SK Hynix, is expected to begin mass manufacturing its response between March or April, enough time ahead for Samsung to begin sending its memory to NVIDIA. And, here, Samsung’s great advantage is that it does not depend on TSMC: it has its own foundry and the HBM4 modules are based on 4 nanometer photolithography.
Looking to the future. SK Hynix’s delay is not because they have rested on their laurels: they are the ones who lead the way in the previous generation thanks to the HBM3E memory, but due to their schedule and because they did not need it, they started developing the new generation later than Samsung. But of course, although HBM is the standard in current AI systems, we have already said that they are expensive chips and, in addition, they heat up a lot, requiring dissipation equipment to match.
And that’s where companies are combining HBM4 memory production with a new generation of DRAM memory. The idea is to find a way for this memory – slower, but cheaper and ‘fresh’ – to compete in bandwidth with the HBM. Samsung and SK Hynix are in it, but they will have to compete against someone who did not play in this league: an Intel that does not arrive alone, but with the help of the Japanese giant SoftBank.

In short: Samsung has proposed to get back on its feet when it comes to manufacturing muscle. And the most important thing of all is that all the companies that make memory modules remain focused on one thing: they make hardware for artificial intelligence while components like RAM and consumer SSDs have sky-high prices.
Images | Maxence Pira, Choi Kwang-mo, NVIDIA logo (edited)
In WorldOfSoftware | Huawei has kept its promise: it has found a way to boost China’s competitiveness in AI compared to the US
