New AI method lets models think harder while avoiding costly bandwidth

New AI method lets models think harder while avoiding costly bandwidth

New AI method lets models think harder while avoiding costly bandwidth


  • DeepSeek’s Engram separates static memory from computation, increasing efficiency in large AI models
  • The method reduces high-speed memory needs by enabling DeepSeek models to use lookups
  • Engram supports asynchronous prefetching across multiple GPUs with minimal performance overhead

DeepSeek, in collaboration with Peking University, introduced a new training method called Engram, designed to decouple memory storage from computational processes.

Traditional large language models require high-bandwidth memory for knowledge retrieval and basic computation, creating a bottleneck in both performance and cost.





Source link

Back To Top