First published: 2016/07/15 (8 years ago) Abstract: Neural networks with recurrent or recursive architecture have shown promising
results on various natural language processing (NLP) tasks. The recurrent and
recursive architectures have their own strength and limitations. The recurrent
networks process input text sequentially and model the conditional transition
between word tokens. In contrast, the recursive networks explicitly model the
compositionality and the recursive structure of natural language. Current
recursive architecture is based on syntactic tree, thus limiting its practical
applicability in different NLP applications. In this paper, we introduce a
class of tree structured model, Neural Tree Indexers (NTI) that provides a
middle ground between the sequential RNNs and the syntactic tree-based
recursive models. NTI constructs a full n-ary tree by processing the input text
with its node function in a bottom-up fashion. Attention mechanism can then be
applied to both structure and different forms of node function. We demonstrated
the effectiveness and the flexibility of a binary-tree model of NTI, showing
the model achieved the state-of-the-art performance on three different NLP
tasks: natural language inference, answer sentence selection, and sentence
classification.