Simple but effective GRU variants

dc.contributor.author Yigit, G.
dc.contributor.author Amasyali, M.F.
dc.date.accessioned 2023-10-19T15:05:33Z
dc.date.available 2023-10-19T15:05:33Z
dc.date.issued 2021
dc.description Kocaeli University;Kocaeli University Technopark en_US
dc.description 2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 --25 August 2021 through 27 August 2021 -- --172175 en_US
dc.description.abstract Recurrent Neural Network (RNN) is a widely used deep learning architecture applied to sequence learning problems. However, it is recognized that RNNs suffer from exploding and vanishing gradient problems that prohibit the early layers of the network from learning the gradient information. GRU networks are particular kinds of recurrent networks that reduce the short-comings of these problems. In this study, we propose two variants of the standard GRU with simple but effective modifications. We applied an empirical approach and tried to determine the effectiveness of the current units and recurrent units of gates by giving different coefficients. Interestingly, we realize that applying such minor and simple changes to the standard GRU provides notable improvements. We comparatively evaluate the standard GRU with the proposed two variants on four different tasks: (1) sentiment classification on the IMDB movie review dataset, (2) language modeling task on Penn TreeBank (PTB) dataset, (3) sequence to sequence addition problem, and (4) question answering problem on Facebook's bAbitasks dataset. The evaluation results indicate that the proposed two variants of GRU consistently outperform standard GRU. © 2021 IEEE. en_US
dc.description.sponsorship ACKNOWLEDGMENT G.Yigit is supported by TUB?TAK - B?DEB 2211/A national fellowship program for Ph.D. studies. en_US
dc.identifier.doi 10.1109/INISTA52262.2021.9548535 en_US
dc.identifier.isbn 9781665436038
dc.identifier.scopus 2-s2.0-85116609087 en_US
dc.identifier.uri https://doi.org/10.1109/INISTA52262.2021.9548535
dc.identifier.uri https://hdl.handle.net/20.500.12469/4942
dc.language.iso en en_US
dc.publisher Institute of Electrical and Electronics Engineers Inc. en_US
dc.relation.ispartof 2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 - Proceedings en_US
dc.rights info:eu-repo/semantics/closedAccess en_US
dc.subject Gated recurrent units en_US
dc.subject Recurrent neural networks en_US
dc.subject Seq2seq en_US
dc.subject Classification (of information) en_US
dc.subject Modeling languages en_US
dc.subject Multilayer neural networks en_US
dc.subject Network layers en_US
dc.subject Gated recurrent unit en_US
dc.subject Gradient informations en_US
dc.subject Learning architectures en_US
dc.subject Learning problem en_US
dc.subject Recurrent networks en_US
dc.subject Seq2seq en_US
dc.subject Sequence learning en_US
dc.subject Short-comings en_US
dc.subject Simple++ en_US
dc.subject Vanishing gradient en_US
dc.subject Recurrent neural networks en_US
dc.title Simple but effective GRU variants en_US
dc.type Conference Object en_US
dspace.entity.type Publication
gdc.author.scopusid 57215312808
gdc.author.scopusid 55664402200
gdc.bip.impulseclass C4
gdc.bip.influenceclass C4
gdc.bip.popularityclass C4
gdc.coar.access metadata only access
gdc.coar.type text::conference output
gdc.collaboration.industrial false
gdc.description.departmenttemp Yigit, G., Kadir Has University, Computer Engineering Department, İstanbul, Turkey; Amasyali, M.F., Yildiz Technical University, Computer Engineering Department, İstanbul, Turkey en_US
gdc.description.endpage 6
gdc.description.publicationcategory Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı en_US
gdc.description.scopusquality N/A
gdc.description.startpage 1
gdc.description.wosquality N/A
gdc.identifier.openalex W3202552216
gdc.index.type Scopus
gdc.oaire.diamondjournal false
gdc.oaire.impulse 10.0
gdc.oaire.influence 4.277298E-9
gdc.oaire.isgreen false
gdc.oaire.keywords Seq2seq
gdc.oaire.keywords Classification (of information)
gdc.oaire.keywords Vanishing gradient
gdc.oaire.keywords Learning problem
gdc.oaire.keywords Simple++
gdc.oaire.keywords Multilayer neural networks
gdc.oaire.keywords Short-comings
gdc.oaire.keywords Gradient informations
gdc.oaire.keywords Recurrent networks
gdc.oaire.keywords Gated recurrent units
gdc.oaire.keywords Network layers
gdc.oaire.keywords Sequence learning
gdc.oaire.keywords Gated recurrent unit
gdc.oaire.keywords Recurrent neural networks
gdc.oaire.keywords Modeling languages
gdc.oaire.keywords Learning architectures
gdc.oaire.popularity 9.818398E-9
gdc.oaire.publicfunded false
gdc.oaire.sciencefields 0301 basic medicine
gdc.oaire.sciencefields 0303 health sciences
gdc.oaire.sciencefields 03 medical and health sciences
gdc.openalex.collaboration National
gdc.openalex.fwci 0.8234871
gdc.openalex.normalizedpercentile 0.84
gdc.opencitations.count 10
gdc.plumx.crossrefcites 1
gdc.plumx.mendeley 25
gdc.plumx.scopuscites 12
gdc.scopus.citedcount 13
gdc.virtual.author Yiğit, Gülsüm
relation.isAuthorOfPublication 363c092e-cd4b-400e-8261-ca5b99b1bea9
relation.isAuthorOfPublication.latestForDiscovery 363c092e-cd4b-400e-8261-ca5b99b1bea9
relation.isOrgUnitOfPublication b20623fc-1264-4244-9847-a4729ca7508c
relation.isOrgUnitOfPublication fd8e65fe-c3b3-4435-9682-6cccb638779c
relation.isOrgUnitOfPublication 2457b9b3-3a3f-4c17-8674-7f874f030d96
relation.isOrgUnitOfPublication.latestForDiscovery b20623fc-1264-4244-9847-a4729ca7508c

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
4942.pdf
Size:
1.2 MB
Format:
Adobe Portable Document Format
Description:
Tam Metin / Full Text