Written by
Nickolay Shmyrev
on
Learning with huge memory
Recently a set of papers were published about "memorization" in neural networks. For example:
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layeralso
Understanding deep learning requires rethinking generalizationIt seems that large memory system has a point, you don’t need millions of computing cores in CPU and, it is too power-expensive, you could just go ahead with very large memory and reasonable amount of cores to access memory with hashing (think of Shazam or
randlm, or
G2P by analogy). You probably do not need heavy tying either.
Advantages are: you can quickly incorporate new knowledge, just put new values in memory, you can model corner cases since they are all still accessible, and, again, you are much more energy-efficient.
Maybe we will see mobile phones with 1Tb of memory sometimes.