diff --git a/examples/cifar10/train_cifar10.py b/examples/cifar10/train_cifar10.py index 24f88e7..9e29963 100644 --- a/examples/cifar10/train_cifar10.py +++ b/examples/cifar10/train_cifar10.py @@ -106,7 +106,7 @@ def train(argv): sched = torch.optim.lr_scheduler.LambdaLR(optim, lr_lambda=warmup_lr) if FLAGS.parallel: print( - "Warning: parallel training is performing slighlty worse than single GPU training due to statistics computation in dataparallel. We recommend to train over a single GPU, which requires around 8 Gb of GPU memory." + "Warning: parallel training is performing slightly worse than single GPU training due to statistics computation in dataparallel. We recommend to train over a single GPU, which requires around 8 Gb of GPU memory." ) net_model = torch.nn.DataParallel(net_model) ema_model = torch.nn.DataParallel(ema_model)