MultiheadAttention — PyTorch 1.10.1 documentation
https://pytorch.org/docs/stable/generated/torch.nn.MultiheadAttention.htmlMultiheadAttention. class torch.nn.MultiheadAttention(embed_dim, num_heads, dropout=0.0, bias=True, add_bias_kv=False, add_zero_attn=False, kdim=None, vdim=None, batch_first=False, device=None, dtype=None) [source] Allows the model to jointly attend to information from different representation subspaces. See Attention Is All You Need.