| import os | |
| def initialize_forge(): | |
| from ldm_patched.modules import args_parser | |
| args_parser.args, _ = args_parser.parser.parse_known_args() | |
| if args_parser.args.gpu_device_id is not None: | |
| os.environ["CUDA_VISIBLE_DEVICES"] = str(args_parser.args.gpu_device_id) | |
| print("Set device to:", args_parser.args.gpu_device_id) | |
| if args_parser.args.cuda_malloc: | |
| from modules_forge.cuda_malloc import try_cuda_malloc | |
| try_cuda_malloc() | |
| import ldm_patched.modules.model_management as model_management | |
| import torch | |
| device = model_management.get_torch_device() | |
| torch.zeros((1, 1)).to(device, torch.float32) | |
| model_management.soft_empty_cache() | |
| import modules_forge.patch_basic | |
| modules_forge.patch_basic.patch_all_basics() | |
| from modules_forge import stream | |
| print("CUDA Stream Activated: ", stream.using_stream) | |
| from modules_forge.shared import diffusers_dir | |
| if "TRANSFORMERS_CACHE" not in os.environ: | |
| os.environ["TRANSFORMERS_CACHE"] = diffusers_dir | |
| if "HF_HOME" not in os.environ: | |
| os.environ["HF_HOME"] = diffusers_dir | |
| if "HF_DATASETS_CACHE" not in os.environ: | |
| os.environ["HF_DATASETS_CACHE"] = diffusers_dir | |
| if "HUGGINGFACE_HUB_CACHE" not in os.environ: | |
| os.environ["HUGGINGFACE_HUB_CACHE"] = diffusers_dir | |
| if "HUGGINGFACE_ASSETS_CACHE" not in os.environ: | |
| os.environ["HUGGINGFACE_ASSETS_CACHE"] = diffusers_dir | |
| if "HF_HUB_CACHE" not in os.environ: | |
| os.environ["HF_HUB_CACHE"] = diffusers_dir | |