debug/torch-distributed-gpu-test.py [60:65]:
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
    with open(__file__, "r") as fh:
        fcntl.flock(fh, fcntl.LOCK_EX)
        try:
            builtins.print(*args, **kwargs)
        finally:
            fcntl.flock(fh, fcntl.LOCK_UN)
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -



