Batch Training RNNs - PyTorch Forums
discuss.pytorch.org › t › batch-training-rnnsMar 08, 2018 · Thanks guys, that did it. I think I just overthought it. I think part of problem was that batch isn’t the first dimension of the input. I will definitely use batch_first=True since that feels way more natural for me. Shame I have to wait 10 more hours until I can implement it, would love to do now. Thanks again, I will post some loss curves here in case it works out. 🙂
Batch Training RNNs - PyTorch Forums
https://discuss.pytorch.org/t/batch-training-rnns/1452507.03.2018 · Hey! If I understand it correctly, when training RNNs using mini batch sgd, the elements in one batch should not be sequential. Rather, every index throughout the batches corresponds to one sequence. I can see that this makes sense when one has multiple sequences to train on. Currently I’m working on a problem where I have only 1 ongoing time series, no …
Batch Training RNNs - PyTorch Forums
discuss.pytorch.org › t › batch-training-rnnsMar 07, 2018 · The cell implementations take one timestep at a time. The LSTM, RNN and GRU all take inputs with several timesteps in one go. I find it helpful to be very clear about the distinction between the batch dimension, whose indices correspond to different input sequences, and the sequence dimension whose indices correspond to different timesteps of each sequence.