“Vanilla” Neural Network
https://duvenaud.github.io/sta414/rnn-slides-stanford.pdfVanilla RNN Gradient Flow h 0 h 1 h 2 h 3 h 4 x 1 x 2 x 3 x 4 Computing gradient of h 0 involves many factors of W (and repeated tanh) Bengio et al, “Learning long-term dependencies with gradient descent is difficult”, IEEE Transactions on Neural Networks, 1994 Pascanu et al, “On the difficulty of training recurrent neural networks ...
A Gentle Introduction to Graph Neural Networks
https://distill.pub/2021/gnn-intro02.09.2021 · A graph is the input, and each component (V,E,U) gets updated by a MLP to produce a new graph. Each function subscript indicates a separate function for a different graph attribute at the n-th layer of a GNN model. As is common with neural networks modules or layers, we can stack these GNN layers together.