Simple but effective GRU variants

Loading...
Publication Logo

Date

2021

Authors

Yigit, G.
Amasyali, M.F.

Journal Title

Journal ISSN

Volume Title

Publisher

Institute of Electrical and Electronics Engineers Inc.

Open Access Color

Green Open Access

No

OpenAIRE Downloads

OpenAIRE Views

Publicly Funded

No
Impulse
Top 10%
Influence
Top 10%
Popularity
Top 10%

Research Projects

Journal Issue

Abstract

Recurrent Neural Network (RNN) is a widely used deep learning architecture applied to sequence learning problems. However, it is recognized that RNNs suffer from exploding and vanishing gradient problems that prohibit the early layers of the network from learning the gradient information. GRU networks are particular kinds of recurrent networks that reduce the short-comings of these problems. In this study, we propose two variants of the standard GRU with simple but effective modifications. We applied an empirical approach and tried to determine the effectiveness of the current units and recurrent units of gates by giving different coefficients. Interestingly, we realize that applying such minor and simple changes to the standard GRU provides notable improvements. We comparatively evaluate the standard GRU with the proposed two variants on four different tasks: (1) sentiment classification on the IMDB movie review dataset, (2) language modeling task on Penn TreeBank (PTB) dataset, (3) sequence to sequence addition problem, and (4) question answering problem on Facebook's bAbitasks dataset. The evaluation results indicate that the proposed two variants of GRU consistently outperform standard GRU. © 2021 IEEE.

Description

Kocaeli University;Kocaeli University Technopark
2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 --25 August 2021 through 27 August 2021 -- --172175

Keywords

Gated recurrent units, Recurrent neural networks, Seq2seq, Classification (of information), Modeling languages, Multilayer neural networks, Network layers, Gated recurrent unit, Gradient informations, Learning architectures, Learning problem, Recurrent networks, Seq2seq, Sequence learning, Short-comings, Simple++, Vanishing gradient, Recurrent neural networks, Seq2seq, Classification (of information), Vanishing gradient, Learning problem, Simple++, Multilayer neural networks, Short-comings, Gradient informations, Recurrent networks, Gated recurrent units, Network layers, Sequence learning, Gated recurrent unit, Recurrent neural networks, Modeling languages, Learning architectures

Fields of Science

0301 basic medicine, 0303 health sciences, 03 medical and health sciences

Citation

WoS Q

N/A

Scopus Q

N/A
OpenCitations Logo
OpenCitations Citation Count
10

Source

2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 - Proceedings

Volume

Issue

Start Page

1

End Page

6
PlumX Metrics
Citations

CrossRef : 1

Scopus : 12

Captures

Mendeley Readers : 25

SCOPUS™ Citations

13

checked on Feb 12, 2026

Page Views

8

checked on Feb 12, 2026

Google Scholar Logo
Google Scholar™
OpenAlex Logo
OpenAlex FWCI
0.8234871

Sustainable Development Goals

SDG data is not available