diff --git a/vllm_hpu_extension/environment.py b/vllm_hpu_extension/environment.py index 79da22b1..27810ec7 100644 --- a/vllm_hpu_extension/environment.py +++ b/vllm_hpu_extension/environment.py @@ -10,8 +10,8 @@ @cache def lazy_logger(): - from vllm.logger import init_logger - return init_logger(__name__) + import logging + return logging.getLogger(__name__) def get_hw(): diff --git a/vllm_hpu_extension/ops.py b/vllm_hpu_extension/ops.py index 36c52ea8..c6b0175b 100644 --- a/vllm_hpu_extension/ops.py +++ b/vllm_hpu_extension/ops.py @@ -11,12 +11,12 @@ import os import torch.nn.functional as F import math +import logging import habana_frameworks.torch.core as htcore -from vllm.logger import init_logger from vllm_hpu_extension.capabilities import capabilities -logger = init_logger(__name__) +logger = logging.getLogger(__name__) HPUFusedRMSNorm = None try: from habana_frameworks.torch.hpex.normalization import FusedRMSNorm diff --git a/vllm_hpu_extension/profiler.py b/vllm_hpu_extension/profiler.py index 68da7b30..460a8643 100644 --- a/vllm_hpu_extension/profiler.py +++ b/vllm_hpu_extension/profiler.py @@ -12,13 +12,13 @@ from typing import Any, List import psutil import torch +import logging from habana_frameworks.torch import torch from vllm_hpu_extension.utils import is_fake_hpu -from vllm.logger import init_logger from vllm.utils import get_vllm_instance_id -logger = init_logger(__name__) +logger = logging.getLogger(__name__) class FileWriter(threading.Thread):