| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import platform |
| | import subprocess |
| | from argparse import ArgumentParser |
| |
|
| | import huggingface_hub |
| |
|
| | from .. import __version__ as version |
| | from ..utils import ( |
| | is_accelerate_available, |
| | is_bitsandbytes_available, |
| | is_flax_available, |
| | is_google_colab, |
| | is_notebook, |
| | is_peft_available, |
| | is_safetensors_available, |
| | is_torch_available, |
| | is_transformers_available, |
| | is_xformers_available, |
| | ) |
| | from . import BaseDiffusersCLICommand |
| |
|
| |
|
| | def info_command_factory(_): |
| | return EnvironmentCommand() |
| |
|
| |
|
| | class EnvironmentCommand(BaseDiffusersCLICommand): |
| | @staticmethod |
| | def register_subcommand(parser: ArgumentParser) -> None: |
| | download_parser = parser.add_parser("env") |
| | download_parser.set_defaults(func=info_command_factory) |
| |
|
| | def run(self) -> dict: |
| | hub_version = huggingface_hub.__version__ |
| |
|
| | safetensors_version = "not installed" |
| | if is_safetensors_available(): |
| | import safetensors |
| |
|
| | safetensors_version = safetensors.__version__ |
| |
|
| | pt_version = "not installed" |
| | pt_cuda_available = "NA" |
| | if is_torch_available(): |
| | import torch |
| |
|
| | pt_version = torch.__version__ |
| | pt_cuda_available = torch.cuda.is_available() |
| |
|
| | flax_version = "not installed" |
| | jax_version = "not installed" |
| | jaxlib_version = "not installed" |
| | jax_backend = "NA" |
| | if is_flax_available(): |
| | import flax |
| | import jax |
| | import jaxlib |
| |
|
| | flax_version = flax.__version__ |
| | jax_version = jax.__version__ |
| | jaxlib_version = jaxlib.__version__ |
| | jax_backend = jax.lib.xla_bridge.get_backend().platform |
| |
|
| | transformers_version = "not installed" |
| | if is_transformers_available(): |
| | import transformers |
| |
|
| | transformers_version = transformers.__version__ |
| |
|
| | accelerate_version = "not installed" |
| | if is_accelerate_available(): |
| | import accelerate |
| |
|
| | accelerate_version = accelerate.__version__ |
| |
|
| | peft_version = "not installed" |
| | if is_peft_available(): |
| | import peft |
| |
|
| | peft_version = peft.__version__ |
| |
|
| | bitsandbytes_version = "not installed" |
| | if is_bitsandbytes_available(): |
| | import bitsandbytes |
| |
|
| | bitsandbytes_version = bitsandbytes.__version__ |
| |
|
| | xformers_version = "not installed" |
| | if is_xformers_available(): |
| | import xformers |
| |
|
| | xformers_version = xformers.__version__ |
| |
|
| | platform_info = platform.platform() |
| |
|
| | is_notebook_str = "Yes" if is_notebook() else "No" |
| |
|
| | is_google_colab_str = "Yes" if is_google_colab() else "No" |
| |
|
| | accelerator = "NA" |
| | if platform.system() in {"Linux", "Windows"}: |
| | try: |
| | sp = subprocess.Popen( |
| | ["nvidia-smi", "--query-gpu=gpu_name,memory.total", "--format=csv,noheader"], |
| | stdout=subprocess.PIPE, |
| | stderr=subprocess.PIPE, |
| | ) |
| | out_str, _ = sp.communicate() |
| | out_str = out_str.decode("utf-8") |
| |
|
| | if len(out_str) > 0: |
| | accelerator = out_str.strip() + " VRAM" |
| | except FileNotFoundError: |
| | pass |
| | elif platform.system() == "Darwin": |
| | try: |
| | sp = subprocess.Popen( |
| | ["system_profiler", "SPDisplaysDataType"], |
| | stdout=subprocess.PIPE, |
| | stderr=subprocess.PIPE, |
| | ) |
| | out_str, _ = sp.communicate() |
| | out_str = out_str.decode("utf-8") |
| |
|
| | start = out_str.find("Chipset Model:") |
| | if start != -1: |
| | start += len("Chipset Model:") |
| | end = out_str.find("\n", start) |
| | accelerator = out_str[start:end].strip() |
| |
|
| | start = out_str.find("VRAM (Total):") |
| | if start != -1: |
| | start += len("VRAM (Total):") |
| | end = out_str.find("\n", start) |
| | accelerator += " VRAM: " + out_str[start:end].strip() |
| | except FileNotFoundError: |
| | pass |
| | else: |
| | print("It seems you are running an unusual OS. Could you fill in the accelerator manually?") |
| |
|
| | info = { |
| | "🤗 Diffusers version": version, |
| | "Platform": platform_info, |
| | "Running on a notebook?": is_notebook_str, |
| | "Running on Google Colab?": is_google_colab_str, |
| | "Python version": platform.python_version(), |
| | "PyTorch version (GPU?)": f"{pt_version} ({pt_cuda_available})", |
| | "Flax version (CPU?/GPU?/TPU?)": f"{flax_version} ({jax_backend})", |
| | "Jax version": jax_version, |
| | "JaxLib version": jaxlib_version, |
| | "Huggingface_hub version": hub_version, |
| | "Transformers version": transformers_version, |
| | "Accelerate version": accelerate_version, |
| | "PEFT version": peft_version, |
| | "Bitsandbytes version": bitsandbytes_version, |
| | "Safetensors version": safetensors_version, |
| | "xFormers version": xformers_version, |
| | "Accelerator": accelerator, |
| | "Using GPU in script?": "<fill in>", |
| | "Using distributed or parallel set-up in script?": "<fill in>", |
| | } |
| |
|
| | print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n") |
| | print(self.format_dict(info)) |
| |
|
| | return info |
| |
|
| | @staticmethod |
| | def format_dict(d: dict) -> str: |
| | return "\n".join([f"- {prop}: {val}" for prop, val in d.items()]) + "\n" |
| |
|