Drastically different inference results ... - discuss.pytorch.org
discuss.pytorch.org › t › drastically-differentFeb 28, 2020 · I trained a segmentation model in Pytorch and tested it to give an F1 score of 0.93 on my local computer (Windows, conda, CUDA 10.2, Pytorch 1.2). However, the F1 score dropped to 0.3 when testing on a Linux server (conda, CUDA 9.0, Pytorch 1.1). I double checked that both sets of code, label files, test sets were the same, and there is no “explicit” random sampling in my code (even if so, the effect shouldn’t be so drastic).