pytorch/rnn.py at master · pytorch/pytorch · GitHub
github.com › pytorch › pytorchwould mean stacking two GRUs together to form a `stacked GRU`, with the second GRU taking in outputs of the first GRU and: computing the final results. Default: 1: bias: If ``False``, then the layer does not use bias weights `b_ih` and `b_hh`. Default: ``True`` batch_first: If ``True``, then the input and output tensors are provided
GRU — PyTorch 1.10.1 documentation
pytorch.org › docs › stableGRU. Applies a multi-layer gated recurrent unit (GRU) RNN to an input sequence. For each element in the input sequence, each layer computes the following function: are the reset, update, and new gates, respectively. * ∗ is the Hadamard product.