TY - GEN
T1 - Evolving memory cell structures for sequence learning
AU - Bayer, Justin
AU - Wierstra, Daan
AU - Togelius, Julian
AU - Schmidhuber, Jürgen
PY - 2009
Y1 - 2009
N2 - Long Short-Term Memory (LSTM) is one of the best recent supervised sequence learning methods. Using gradient descent, it trains memory cells represented as differentiable computational graph structures. Interestingly, LSTM's cell structure seems somewhat arbitrary. In this paper we optimize its computational structure using a multi-objective evolutionary algorithm. The fitness function reflects the structure's usefulness for learning various formal languages. The evolved cells help to understand crucial features that aid sequence learning.
AB - Long Short-Term Memory (LSTM) is one of the best recent supervised sequence learning methods. Using gradient descent, it trains memory cells represented as differentiable computational graph structures. Interestingly, LSTM's cell structure seems somewhat arbitrary. In this paper we optimize its computational structure using a multi-objective evolutionary algorithm. The fitness function reflects the structure's usefulness for learning various formal languages. The evolved cells help to understand crucial features that aid sequence learning.
UR - http://www.scopus.com/inward/record.url?scp=70450190492&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=70450190492&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-04277-5_76
DO - 10.1007/978-3-642-04277-5_76
M3 - Conference contribution
AN - SCOPUS:70450190492
SN - 3642042767
SN - 9783642042768
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 755
EP - 764
BT - Artificial Neural Networks - ICANN 2009 - 19th International Conference, Proceedings
T2 - 19th International Conference on Artificial Neural Networks, ICANN 2009
Y2 - 14 September 2009 through 17 September 2009
ER -