By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: Tensormesh raises $4.5M to squeeze more inference out of AI server loads | News
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > Tensormesh raises $4.5M to squeeze more inference out of AI server loads | News
News

Tensormesh raises $4.5M to squeeze more inference out of AI server loads | News

News Room
Last updated: 2025/10/23 at 11:18 AM
News Room Published 23 October 2025
Share
SHARE

With the AI infrastructure push reaching staggering proportions, there’s more pressure than ever to squeeze as much inference as possible out of the GPUs they have. And for researchers with expertise in a particular technique, it’s a great time to raise funding.

That’s part of the driving force behind Tensormesh, launching out of stealth this week with $4.5 million in seed funding. The investment was led by Laude Ventures, with additional angel funding from database pioneer Michael Franklin.

Tensormesh is using the money to build a commercial version of the open-source LMCache utility, launched and maintained by Tensormesh co-founder Yihua Cheng. Used well, LMCache can reduce inference costs by as much as ten times — a power that’s made it a staple in open-source deployments and drawn in integrations from heavy-hitters like Google and Nvidia. Now, Tensormesh is planning to parlay that academic reputation into a viable business.

The heart of the key-value cache (or KV cache), a memory system used to process complex inputs more efficiently by condensing them down to their key values. In traditional architectures, the KV cache is discarded at the end of each query — but TensorMesh CEO Juchen Jiang argues that this is an enormous source of inefficiency.

“It’s like having a very smart analyst reading all the data, but they forget what they have learned after each question,” says Tensormesh co-founder Junchen Jiang.

Instead of discarding that cache, Tensormesh’s systems hold onto it, allowing it to be redeployed when the model executes a similar process in a separate query. Because GPU memory is so precious, this can mean spreading data across several different storage layers, but the reward is significantly more inference power for the same server load.

The change is particularly powerful for chat interfaces, since models need to continually refer back to the growing chat log as the conversation progresses. Agentic systems have a similar issue, with a growing log of actions and goals.

In theory, these are changes AI companies can execute on their own — but the technical complexity makes it a daunting task. Given the Tensormesh team’s work researching the process and the intricacy of the detail itself, the company is betting there will be lots of demand for an out-of-the-box product.

“Keeping the KV cache in a secondary storage system and reused efficiently without slowing the whole system down is a very challenging problem,” says Jiang. “We’ve seen people hire 20 engineers and spend three or four months to build such a system. Or they can use our product and do it very efficiently.”

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article Reddit accuses Perplexity of stealing content to train AI
Next Article Social media for nonprofits: 14 expert-backed tips for success
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

AirPods Pro 3 on Android review: great sound, frustrating features
News
Tesla recalls over 63,000 Cybertrucks due to the front lights being too bright
Software
Veeam to buy data security strategy management company Securiti AI
Mobile
Meet Contiant, Picky, and Dar Blockchain: HackerNoon Startups of the Week | HackerNoon
Computing

You Might also Like

News

AirPods Pro 3 on Android review: great sound, frustrating features

7 Min Read
News

45% of AI-generated news is wrong, new study warns — here’s what happened when I tested it myself

7 Min Read

Rare dinosaur mummies help scientists recreate their prehistoric lives

3 Min Read
News

Gemini Comes To GM Vehicles As CarPlay And Android Auto Are Phased Out – BGR

6 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?