This Medium
post is unfortunately member only, but the title is impressive: "LLMs and Memory is Definitely All You Need: Google Shows that Memory-Augmented LLMs Can Simulate Any Turing Machine"
I think this is another example of one point of Anatoly: nobody will insist on pure LLM if they found a new useful feature.
Abstract
We show that transformer-based large language models are computationally universal when augmented with an external memory. Any deterministic language model that conditions on strings of bounded length is equivalent to a finite automaton, hence computationally limited. However, augmenting such models with a read-write memory creates the possibility of processing arbitrarily large inputs and, potentially, simulating any algorithm. We establish that an existing large language model, Flan-U-PaLM 540B, can be combined with an associative read-write memory to exactly simulate the execution of a universal Turing machine, U15,2. A key aspect of the finding is that it does not require any modification of the language model weights. Instead, the construction relies solely on designing a form of stored instruction computer that can subsequently be programmed with a specific set of prompts.
Alex
вторник, 25 июля 2023 г. в 15:50:33 UTC+3, alex.shkotin: