Lo nuevo en Article Factory y lo último en el mundo de la IA generativa

DeepSeek Introduces Engram to Cut High‑Bandwidth Memory Needs in Large AI Models

DeepSeek Introduces Engram to Cut High‑Bandwidth Memory Needs in Large AI Models
DeepSeek, in partnership with Peking University, unveiled Engram, a new training method that separates static memory from computation in large language models. By using hashed N‑gram lookups and a context‑aware gating mechanism, Engram reduces reliance on high‑bandwidth memory (HBM), allowing models to operate efficiently on standard GPU memory while scaling parameter counts. Tests on a 27‑billion‑parameter model showed measurable gains across industry benchmarks, and the approach integrates with existing hardware solutions such as Phison’s SSD‑based accelerators and emerging CXL standards. Engram could ease pressure on costly memory hardware and stabilize DRAM price volatility. Leer más →