Embedding — PyTorch 1.10.1 documentation
pytorch.org › generated › torchFor a newly constructed Embedding, the embedding vector at padding_idx will default to all zeros, but can be updated to another value to be used as the padding vector. max_norm (float, optional) – If given, each embedding vector with norm larger than max_norm is renormalized to have norm max_norm.