[Misc] Add logger.exception for TPU information collection failures (#20510)

Signed-off-by: reidliu41 <reid201711@gmail.com>
This commit is contained in:
Reid
2025-07-05 22:24:32 +08:00
committed by GitHub
parent 32c9be2200
commit cf4cd53982

View File

@ -20,9 +20,12 @@ import torch
import vllm.envs as envs
from vllm.connections import global_http_connection
from vllm.logger import init_logger
from vllm.utils import cuda_device_count_stateless, cuda_get_device_properties
from vllm.version import __version__ as VLLM_VERSION
logger = init_logger(__name__)
_config_home = envs.VLLM_CONFIG_ROOT
_USAGE_STATS_JSON_PATH = os.path.join(_config_home, "usage_stats.json")
_USAGE_STATS_DO_NOT_TRACK_PATH = os.path.join(_config_home, "do_not_track")
@ -183,7 +186,7 @@ class UsageMessage:
self.gpu_memory_per_device = (
torch_xla.core.xla_model.get_memory_info()["bytes_limit"])
except Exception:
pass
logger.exception("Failed to collect TPU information")
self.provider = _detect_cloud_provider()
self.architecture = platform.machine()
self.platform = platform.platform()