|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import importlib
|
|
import sys
|
|
|
|
|
|
if sys.version_info < (3, 8):
|
|
_is_python_greater_3_8 = False
|
|
else:
|
|
_is_python_greater_3_8 = True
|
|
|
|
|
|
def is_peft_available() -> bool:
|
|
return importlib.util.find_spec("peft") is not None
|
|
|
|
|
|
def is_unsloth_available() -> bool:
|
|
return importlib.util.find_spec("unsloth") is not None
|
|
|
|
|
|
def is_accelerate_greater_20_0() -> bool:
|
|
if _is_python_greater_3_8:
|
|
from importlib.metadata import version
|
|
|
|
accelerate_version = version("accelerate")
|
|
else:
|
|
import pkg_resources
|
|
|
|
accelerate_version = pkg_resources.get_distribution("accelerate").version
|
|
return accelerate_version >= "0.20.0"
|
|
|
|
|
|
def is_transformers_greater_than(version: str) -> bool:
|
|
_transformers_version = importlib.metadata.version("transformers")
|
|
return _transformers_version > version
|
|
|
|
|
|
def is_torch_greater_2_0() -> bool:
|
|
if _is_python_greater_3_8:
|
|
from importlib.metadata import version
|
|
|
|
torch_version = version("torch")
|
|
else:
|
|
import pkg_resources
|
|
|
|
torch_version = pkg_resources.get_distribution("torch").version
|
|
return torch_version >= "2.0"
|
|
|
|
|
|
def is_diffusers_available() -> bool:
|
|
return importlib.util.find_spec("diffusers") is not None
|
|
|
|
|
|
def is_bitsandbytes_available() -> bool:
|
|
import torch
|
|
|
|
|
|
return importlib.util.find_spec("bitsandbytes") is not None and torch.cuda.is_available()
|
|
|
|
|
|
def is_torchvision_available() -> bool:
|
|
return importlib.util.find_spec("torchvision") is not None
|
|
|
|
|
|
def is_rich_available() -> bool:
|
|
return importlib.util.find_spec("rich") is not None
|
|
|
|
|
|
def is_wandb_available() -> bool:
|
|
return importlib.util.find_spec("wandb") is not None
|
|
|
|
|
|
def is_xpu_available() -> bool:
|
|
if is_accelerate_greater_20_0():
|
|
import accelerate
|
|
|
|
return accelerate.utils.is_xpu_available()
|
|
else:
|
|
if importlib.util.find_spec("intel_extension_for_pytorch") is None:
|
|
return False
|
|
try:
|
|
import torch
|
|
|
|
return hasattr(torch, "xpu") and torch.xpu.is_available()
|
|
except RuntimeError:
|
|
return False
|
|
|
|
|
|
def is_npu_available() -> bool:
|
|
"""Checks if `torch_npu` is installed and potentially if a NPU is in the environment"""
|
|
if importlib.util.find_spec("torch") is None or importlib.util.find_spec("torch_npu") is None:
|
|
return False
|
|
|
|
import torch
|
|
import torch_npu
|
|
|
|
return hasattr(torch, "npu") and torch.npu.is_available()
|
|
|