{"id":48673,"date":"2026-01-13T12:50:22","date_gmt":"2026-01-13T04:50:22","guid":{"rendered":"https:\/\/www.1ai.net\/?p=48673"},"modified":"2026-01-13T12:50:22","modified_gmt":"2026-01-13T04:50:22","slug":"%e6%a2%81%e6%96%87%e9%94%8b%e7%bd%b2%e5%90%8d%e6%96%b0%e8%ae%ba%e6%96%87%e6%9b%9d%e5%85%89%ef%bc%9adeepseek-v4-%e6%88%96%e5%bc%95%e5%85%a5%e5%85%a8%e6%96%b0%e8%ae%b0%e5%bf%86%e6%9e%b6%e6%9e%84","status":"publish","type":"post","link":"https:\/\/www.1ai.net\/en\/48673.html","title":{"rendered":"Leung Wensai's new paper came to light: DeepSeek V4 or introduced a new memory structure"},"content":{"rendered":"<p>The news of January 13, this morning<a href=\"https:\/\/www.1ai.net\/en\/tag\/deepseek\" title=\"[View articles tagged with [DeepSeek]]\" target=\"_blank\" >DeepSeek<\/a> Open source complete new architecture module \"Engram\" and synchronized the release of technical papers, re-emerged in the author<a href=\"https:\/\/www.1ai.net\/en\/tag\/%e6%a2%81%e6%96%87%e9%94%8b\" title=\"[See articles with labels]\" target=\"_blank\" >Leung Man Fung<\/a>.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-48674\" title=\"3caef47dj00t8sdf1002hd000ujm\" src=\"https:\/\/www.1ai.net\/wp-content\/uploads\/2026\/01\/3caef47dj00t8sdf1002hd000u000jcm.jpg\" alt=\"3caef47dj00t8sdf1002hd000ujm\" width=\"1080\" height=\"696\" \/><\/p>\n<p>It has been learned that the Engram module, by introducing a scalable searchable memory structure, provides a completely new slender dimension for the larger model, different from the traditional Transformer and MoE\u3002<\/p>\n<p>In a paper, DeepSeek noted that the current mainstream large model was structurally inefficient in dealing with two types of tasks: One is the \"table\" memory that relies on fixed knowledge, and the other is complex reasoning and combination calculations\u3002<\/p>\n<p>The traditional Transformer (whether Dense or MoE) has to re-establish these static patterns through multi-layered attention and MLP, leading to a significant consumption of computing resources on \"repeated construction of known models\"\u3002<\/p>\n<p>Engram's core mechanism is based on modern Hashi N-gram embedded O(1) search memory. The module will perform N-gram slices for input token sequences and achieves constant-time retrieval through multi-Hashi mapping to an extended static memory table\u3002<\/p>\n<p>It was stressed that such searches were not related to the size of the model and that the search costs remained stable even if the memory tables were extended to a billion-scale parameter\u3002<\/p>\n<p>In contrast to MoE's calculations, Engram offers \"conditional memory\". The module will determine whether the search results will be enabled according to the current context and will be integrated with the backbone network through a door-control mechanism\u3002<\/p>\n<p>The paper showed that Engram was usually placed in the early stages of the model to take on the role of \"model reconstruction\", thus releasing the depth of the calculation of the subsequent layer for complex reasoning\u3002<\/p>\n<p>DeepSeek, in an experiment of 27B parameter sizes, redistributed part of the MoE expert parameter to the Engram memory table, and the model was significantly upgraded in terms of knowledge, reasoning, code and mathematical tasks under the same parameters and equal computational conditions\u3002<\/p>\n<p>On the X platform, the technical discussions concluded that the Engram mechanism had been effective in reducing the need for re-establishment of static models at the early stages of the model, making the model more \"deep\" in the reasoning part\u3002<\/p>\n<p>SOME DEVELOPERS POINT OUT THAT THIS STRUCTURE ALLOWS LARGE-SCALE STATIC MEMORY TO BE REMOVED FROM THE GPU STORAGE LIMIT AND TO PREFEASIBILITY OF HOST MEMORY THROUGH A DEFINITIVE LOCATION, THUS KEEPING COSTS LOW AT THE REASONING STAGE\u3002<\/p>\n<p>Many observers speculate that Engram is likely to be the core technology base for the next generation of DeepSeek models, V4\u3002<\/p>","protected":false},"excerpt":{"rendered":"<p>On January 13th, in the early hours of this morning, DeepSeek opened a new architecture module, \"Engram\", and published a technical paper simultaneously, which was re-emerged in the author's name. It has been learned that the Engram module, by introducing a scalable searchable memory structure, provides a completely new slender dimension for the larger model, different from the traditional Transformer and MoE. DeepSeek notes in his paper that the current mainstream large model is structurally inefficient in dealing with two types of tasks: One is \"table\" memory that relies on fixed knowledge, and the other is complex reasoning and combination calculations. Traditional Transformer (whether Dense or MoE) needs to recreate these static patterns through multi-layered attention and MLP<\/p>","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[146],"tags":[3606,5762],"collection":[],"class_list":["post-48673","post","type-post","status-publish","format-standard","hentry","category-news","tag-deepseek","tag-5762"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/48673","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/comments?post=48673"}],"version-history":[{"count":0,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/posts\/48673\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/media?parent=48673"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/categories?post=48673"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/tags?post=48673"},{"taxonomy":"collection","embeddable":true,"href":"https:\/\/www.1ai.net\/en\/wp-json\/wp\/v2\/collection?post=48673"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}