2006.15704_PyTorch Distributed: Experiences on Accelerating Data Parallel Training