MIT researchers developed Attention Matching, a KV cache compaction technique that compresses LLM memory by 50x in seconds — ...
Huawei Data Storage officially launched a "3+1" AI Data Platform to tackle these challenges: Knowledge generation and retrieval with high-accuracy multimodal knowledge for more accurate retrieval ...
Micron is pushing a new SOCAMM2 module that targets the memory choke points showing up in long-context AI workloads. The outfit claims it has set a “new benchmark” by raising per-module capacity to ...
At MWC Barcelona 2026 the president of Huawei Data Storage Product Line shared Huawei's key insights and innovations ...
BARCELONA, Spain, March 5, 2026 /PRNewswire/ -- At the Huawei AI DC Innovation Forum at MWC Barcelona 2026, Huawei unveiled its AI Data Platform, designed to address the key challenges in adopting AI ...
CHATSWORTH, Calif. — July 18, 2025 DDN today unveiled performance benchmarks that the company said demonstrates how its AI-optimized DDN Infinia platform eliminates GPU waste and delivers the fastest ...
A new technical paper titled “Accelerating LLM Inference via Dynamic KV Cache Placement in Heterogeneous Memory System” was published by researchers at Rensselaer Polytechnic Institute and IBM. “Large ...