in bench_cluster/communication/utils.py [0:0]
def sync_all(): torch.cuda.synchronize() dist.barrier()