A Gated Recurrent Unit, or GRU, is a type of recurrent neural network. It is similar to an LSTM, but only has two gates - a reset gate and an update gate - and notably lacks an output gate. Fewer parameters means GRUs are generally easier/faster to train than their LSTM counterparts.
Image Source: here
Source: Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine TranslationPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Time Series Analysis | 51 | 6.58% |
Speech Synthesis | 40 | 5.16% |
Language Modelling | 25 | 3.23% |
Sentence | 25 | 3.23% |
Sentiment Analysis | 20 | 2.58% |
Time Series Forecasting | 19 | 2.45% |
General Classification | 19 | 2.45% |
Classification | 17 | 2.19% |
Text-To-Speech Synthesis | 14 | 1.81% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |