By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: How much information really memorizes a LLM?
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > Mobile > How much information really memorizes a LLM?
Mobile

How much information really memorizes a LLM?

News Room
Last updated: 2025/06/06 at 5:51 PM
News Room Published 6 June 2025
Share
SHARE

The debate between memorization and generalization in the LLM (extensive language models) is not new. But it is rarely addressed with the almost surgical precision with which a group of target researchers, Google Deepmind, Nvidia and the University of Cornell have done so. How much can a Llm of your training data remember? When do you stop memorizing concrete facts and start learning patterns? These questions, apparently philosophical, are actually quantifiable. And that is just what has been achieved in this new study: measure, exactly, how many bits a model can store.

The border between the memorized and the learned in the LLM is not only diffuse: is invisible to the naked eye. Often, a correct answer does not necessarily mean that the model has understood the concept, but could have stored word by word. The work of these researchers seeks precisely to draw a line between both phenomena. To do this, they resorted to a meticulous strategy: train hundreds of language models from scratch, using both synthetic and real datasets, carefully deduced to avoid involuntary repetitions. The sizes of the models ranged between 500,000 and 1.5 billion parameters, with architectures similar to those of GPT.

The most striking innovation of the study is a metric called HK, based on the complexity of Kolmogorov, which It allows estimating how much original and specific information has been really stored by the model. And with this tool, the team obtained a key fact: a MMORIZA, on average, between 3.5 and 3.6 bits for each of its parameters. That is the maximum capacity before the “sature” model its space and begins to generalize, abandoning literal repetition to embrace broader statistical patterns.

This transition is not simply theoretical: it manifests itself in the form of a double fall in the validation error, the well -known phenomenon of the Double Descent, which marks the moment when the model begins to behave more generalized. Upon reaching the saturation threshold, the memorized data stop providing additional value and the model reorganizes its internal “memory” to optimize learning.

Another relevant observation is that the type of precision with which the model is trained – such as BFloat16 compared to FP32 – has a minimum effect on this memorization capacity. Instead, the researchers found that The most prone examples to be memorized were those with rare tokenssyntactically rare sequences or phrases in minority languages. That is, the model tends to store the unusual, which moves away from the dominant pattern.

This type of memorization not only has technical implications. It also raises issues about privacy, audit and traceability. If a model memorizes rare examples that contain sensitive information, it could reproduce them without their designers knowing it. Understanding this quantitative limit becomes, then, a critical tool to evaluate the real behavior of the LLM.

The study does not intend to reduce the size of the models or directly improve their computational efficiency. Your contribution is on another plane: better understand how and how much can “remember” a model. A contribution that also feels the basis for future research on how to control, limit or even audit that memorization process.

Perhaps the most valuable of this work is that, by measuring what seemed immersible, manages to return some transparency to a often treated land like a black box. At a time when LLMs are already part of the infrastructure of our digital life, knowing how much they remember is not an academic curiosity, but an urgent need. Because only understanding what is within a model, we can trust what comes out of it.

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Google Gemini can now handle scheduled tasks like an assistant
Next Article Perplexity Ai Ceo Pans Google’s Ai Rollout and Assistant Product
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

Mistral’s New Vibe Coding Platform to Compete With Github Copilot
Software
Atlantic Starr singer dies aged 68 as fans pay tribute to ‘voice of a generation
News
Gen Z Consumer Behavior & How It Impacts Your Campaigns
Computing
Which Samsung Galaxy Phone Should You Buy?
Gadget

You Might also Like

Mobile

LG Ultrafine 40WT95UF, Premium Monitor for Companies

3 Min Read
Mobile

Borderlands 2 and Deathloop Are Free to Claim Right Now

2 Min Read
Mobile

What cars can circulate and which ones rest on June 7

4 Min Read
Mobile

The Wacom Cintiq are renewed with three new models

3 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?