mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-21 05:34:18 +08:00
[BE]: Use proper logger in torch.distributed.run (#140547)
`torch.distributed.run` was improperly using the root logger and ignoring all logging settings and useful debugging info. Now properly uses the correct logger. Will be added to ruff as part of LOG015 soon. Pull Request resolved: https://github.com/pytorch/pytorch/pull/140547 Approved by: https://github.com/XuehaiPan, https://github.com/fegin
This commit is contained in:
committed by
PyTorch MergeBot
parent
1c669e7c4e
commit
72c6d13cea
@ -393,7 +393,6 @@ utility
|
||||
main()
|
||||
|
||||
"""
|
||||
import logging
|
||||
import os
|
||||
import sys
|
||||
import uuid
|
||||
@ -673,7 +672,7 @@ def parse_min_max_nnodes(nnodes: str):
|
||||
|
||||
def determine_local_world_size(nproc_per_node: str):
|
||||
try:
|
||||
logging.info("Using nproc_per_node=%s.", nproc_per_node)
|
||||
logger.info("Using nproc_per_node=%s.", nproc_per_node)
|
||||
return int(nproc_per_node)
|
||||
except ValueError as e:
|
||||
if nproc_per_node == "cpu":
|
||||
@ -762,7 +761,7 @@ def _get_logs_specs_class(logs_specs_name: Optional[str]) -> Type[LogsSpecs]:
|
||||
f"Could not find entrypoint under 'torchrun.logs_specs[{logs_specs_name}]' key"
|
||||
)
|
||||
|
||||
logging.info(
|
||||
logger.info(
|
||||
"Using logs_spec '%s' mapped to %s", logs_specs_name, str(logs_specs_cls)
|
||||
)
|
||||
else:
|
||||
|
Reference in New Issue
Block a user