This article serves to briefly outline the design of the Neural Turing Machine (NTM), a backpropogatable architecture that can (among many possibilities) learn to dynamically execute programs.
I’ve added some specifications about the NTM’s architecture that the paper excludes for the sake of generality. These will be discussed upon presentation.
The Neural Turing Machine was proposed by Graves et al. as a Turing-Complete network capable of learning (rather complex) programs. Inspired by the sequential nature of the brain, and the large, addressable memory of the computer.
The controller acts as the interface between input data and memory. It learns to manage its own memory through addressing.
The paper maintains that the controller can be of any format, it simply needs to read in data and produce the outputs required by the sub-modules. They choose an LSTM for their implementation.
This module generates a window over the memory for the read and write heads.
This module effectively laps over the memory comparing each block to the key vector and creates a normally distributed weighting over the memory based of the similarity.
where is an example difference function.
This module has 3 steps:
Writing is performed in two steps, similar to how an LSTM updates its state.
This architecture has already had an impact on a multitude of research projects (notably, the Neural GPU) and I have great faith it will continue to do so. There’s been much discussion over the past couple of years about computers programming themselves, and I believe this has been the greatest stride towards that end-goal.
The Neural Turing Machine is a reactionary computer, changing behaviour based on its “environment”. It certainly will play a major role in setting precedence for the way neural networks will be applied for the purpose of AI.
I look forward to reading more of the work done by the Google DeepMind and OpenAI teams.
If there are any errors in my description please do not hesistate to reach out to me at firstname.lastname@example.org.