North Coast Synthesis Ltd.

Table lookups again, with Engram

◀ Prev | 2026-03-30, access: $ Basic

Popular techniques in language modelling, including RAG, MoE, and attention itself, amount to replacing as much as possible of a neural network model with different kinds of table lookups.  In this recent paper from DeepSeek's research group, they attempt another such replacement:  shifting factual knowledge out of the model weights as such, into a separate hash table.

Video DeepSeek MoE RAG text Popular techniques in language modelling, including RAG, MoE, and attention itself, amount to replacing as much as possible of a neural network model with different kinds of table lookups. In this recent paper from DeepSeek's research group, they attempt another such replacement: shifting factual knowledge out of the model weights as such, into a separate hash table.