A Comparison of LSTM and GRU Networks for Learning Symbolic Sequences

Research output: Chapter in Book/Conference proceedingConference contributionpeer-review

Abstract

We explore the architecture of recurrent neural networks (RNNs) by studying the complexity of string sequences that it is able to memorize. Symbolic sequences of different complexity are generated to simulate RNN training and study parameter configurations with a view to the network’s capability of learning and inference. We compare Long Short-Term Memory (LSTM) networks and gated recurrent units (GRUs). We find that an increase in RNN depth does not necessarily result in better memorization capability when the training time is constrained. Our results also indicate that the learning rate and the number of units per layer are among the most important hyper-parameters to be tuned. Generally, GRUs outperform LSTM networks on low-complexity sequences while on high-complexity sequences LSTMs perform better.
Original languageEnglish
Title of host publicationIntelligent Computing
Subtitle of host publicationProceedings of the 2023 Computing Conference
EditorsKohei Arai
Place of PublicationCham
PublisherSpringer Nature Switzerland AG
Pages771-785
Number of pages15
Volume2
ISBN (Electronic)9783031379635
ISBN (Print)9783031379628
Publication statusPublished - 12 Sept 2023

Publication series

NameLecture Notes in Networks and Systems
PublisherSpringer
Volume739
ISSN (Print)2367-3370
ISSN (Electronic)2367-3389

Fingerprint

Dive into the research topics of 'A Comparison of LSTM and GRU Networks for Learning Symbolic Sequences'. Together they form a unique fingerprint.

Cite this