TY - JOUR
T1 - Tree-structured composition in neural networks without tree-structured architectures
AU - Bowman, Samuel R.
AU - Manning, Christopher D.
AU - Potts, Christopher
N1 - Funding Information:
We gratefully acknowledge a Google Faculty Research Award, a gift from Bloomberg L.P., and support from the Defense Advanced Research Projects Agency (DARPA) Deep Exploration and Fil-tering of Text (DEFT) Program under Air Force Research Laboratory (AFRL) contract no. FA8750- 13-2-0040, the National Science Foundation under grant no. IIS 1159679, and the Department of the Navy, Office of Naval Research, under grant no. N00014-13-1-0287. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of Google, Bloomberg L.P., DARPA, AFRL, NSF, ONR, or the US government.
Publisher Copyright:
Copyright © 2015 for this paper by its authors.
PY - 2015
Y1 - 2015
N2 - Tree-structured neural networks encode a particular tree geometry for a sentence in the network design. However, these models have at best only slightly outperformed simpler sequence-based models. We hypothesize that neural sequence models like LSTMs are in fact able to discover and implicitly use recursive compositional structure, at least for tasks with clear cues to that structure in the data. We demonstrate this possibility using an artificial data task for which recursive compositional structure is crucial, and find an LSTM-based sequence model can indeed learn to exploit the underlying tree structure. However, its performance consistently lags behind that of tree models, even on large training sets, suggesting that tree-structured models are more effective at exploiting recursive structure.
AB - Tree-structured neural networks encode a particular tree geometry for a sentence in the network design. However, these models have at best only slightly outperformed simpler sequence-based models. We hypothesize that neural sequence models like LSTMs are in fact able to discover and implicitly use recursive compositional structure, at least for tasks with clear cues to that structure in the data. We demonstrate this possibility using an artificial data task for which recursive compositional structure is crucial, and find an LSTM-based sequence model can indeed learn to exploit the underlying tree structure. However, its performance consistently lags behind that of tree models, even on large training sets, suggesting that tree-structured models are more effective at exploiting recursive structure.
UR - http://www.scopus.com/inward/record.url?scp=84977505088&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84977505088&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:84977505088
SN - 1613-0073
VL - 1583
JO - CEUR Workshop Proceedings
JF - CEUR Workshop Proceedings
T2 - NIPS Workshop on Cognitive Computation, CoCo 2015
Y2 - 11 December 2015 through 12 December 2015
ER -