Ask Me Anything: Dynamic Memory Networks for Natural Language Processing

Goal

a framework for general question answering tasks trained using (input, question, answer).

Tasks

sequence tagging, classification, sequence-to-sequence(seq2seq), question answering(QA) that requires transiitve reasoning.

Model

model

Input Module

encode raw inputs(images, texts(sentences, story, tweets…etc) ) to distibuted vector representation.

GRU is used instead of LSTM for computation efficiency with similar performence.

Question Module

encode raw question of tasks. Then fed into espsodic memory module(EMM) as initial state/basis for it(EMM) to iterates.

Share word embedding matrix accross Input module and Question Module.

Episodic Memory Module

Iteratively attend different part of the input** given the current state.

(share some similarity to hops between controller module and memory module from End-To-End Memory Networks(MemN2N))

Compromised of an attention mechanism within a recurrent network.

Score function G of the answer is an 2-layer feed forward Neural Network as well.

Criteria for stopping includes a end-of-passes representation for the GRU to choose as well as a maximum number of iteration just in case.

Answer Module

Generate answer from final memory vector from EMM.

model_example

Training

Experiments

###QA

QA

TODO: Experiments is not complete