MIT researchers developed Attention Matching, a KV cache compaction technique that compresses LLM memory by 50x in seconds — ...
Huawei Data Storage officially launched a "3+1" AI Data Platform to tackle these challenges: Knowledge generation and retrieval with high-accuracy multimodal knowledge for more accurate retrieval ...
Micron is pushing a new SOCAMM2 module that targets the memory choke points showing up in long-context AI workloads. The outfit claims it has set a “new benchmark” by raising per-module capacity to ...
At MWC Barcelona 2026 the president of Huawei Data Storage Product Line shared Huawei's key insights and innovations ...
BARCELONA, Spain, March 5, 2026 /PRNewswire/ -- At the Huawei AI DC Innovation Forum at MWC Barcelona 2026, Huawei unveiled its AI Data Platform, designed to address the key challenges in adopting AI ...
CHATSWORTH, Calif. — July 18, 2025 DDN today unveiled performance benchmarks that the company said demonstrates how its AI-optimized DDN Infinia platform eliminates GPU waste and delivers the fastest ...
A new technical paper titled “Accelerating LLM Inference via Dynamic KV Cache Placement in Heterogeneous Memory System” was published by researchers at Rensselaer Polytechnic Institute and IBM. “Large ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results