[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

PEAK:AIO Solves Long-Running AI Memory Bottleneck for LLM Inference and Model Innovation with Unified Token Memory Feature

logo

PEAK:AIO, the data infrastructure pioneer redefining AI-first data acceleration, today unveiled the first dedicated solution to unify KVCache acceleration and GPU memory expansion for large-scale AI workloads, including inference, agentic systems, and model creation.

As AI workloads evolve beyond static prompts into dynamic context streams, model creation pipelines, and long-running agents, infrastructure must evolve, too.

Also Read: The Impact of Increased AI Investment on Organizational AI Strategies

“Whether you are deploying agents that think across sessions or scaling toward million-token context windows, where memory demands can exceed 500GB per model, this appliance makes it possible by treating token history as memory, not storage,” said Eyal Lemberger, Chief AI Strategist and Co-Founder of PEAK:AIO “It is time for memory to scale like compute has.”

As transformer models grow in size and context, AI pipelines face two critical limitations: KVCache inefficiency and GPU memory saturation. Until now, vendors have retrofitted legacy storage stacks or overextended NVMe to delay the inevitable. PEAK:AIO’s new 1U Token Memory Feature changes that by building for memory, not files.

The First Token-Centric Architecture Built for Scalable AI

Related Posts
1 of 42,265

Powered by CXL memory and integrated with Gen5 NVMe and GPUDirect RDMA, PEAK:AIO’s feature delivers up to 150 GB/sec sustained throughput with sub-5 microsecond latency. It enables:

  • KVCache reuse across sessions, models, and nodes
  • Context-window expansion for longer LLM history
  • GPU memory offload via true CXL tiering
  • Ultra-low latency access using RDMA over NVMe-oF

This is the first feature that treats token memory as infrastructure rather than storage, allowing teams to cache token history, attention maps, and streaming data at memory-class latency.

Unlike passive NVMe-based storage, PEAK:AIO’s architecture aligns directly with NVIDIA’s KVCache reuse and memory reclaim models. This provides plug-in support for teams building on TensorRT-LLM or Triton, accelerating inference with minimal integration effort. By harnessing true CXL memory-class performance, it delivers what others cannot: token memory that behaves like RAM, not files.

Also Read: The Evolution of Data Engineering: Making Data AI-Ready

“While others are bending file systems to act like memory, we built infrastructure that behaves like memory, because that is what modern AI needs,” continued  Lemberger. “At scale, it is not about saving files; it is about keeping every token accessible in microseconds. That is a memory problem, and we solved it at embracing the latest silicon layer.”

The fully software-defined solution utilizes off-the-shelf servers is expected to enter production by Q3. To discuss early access, technical consultation, or how PEAK:AIO can support AI infrastructure needs,

[To share your insights with us, please write to psen@itechseries.com]

Comments are closed.