code
stringlengths
114
1.05M
path
stringlengths
3
312
quality_prob
float64
0.5
0.99
learning_prob
float64
0.2
1
filename
stringlengths
3
168
kind
stringclasses
1 value
import re from .compat import unicode class Pattern(object): """ The :class:`Pattern` class is the abstract definition of a pattern. """ # Make the class dict-less. __slots__ = ("include",) def __init__(self, include): """ Initializes the :class:`Pattern` instance. ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/thirdparty/pathspec/pattern.py
0.894138
0.489626
pattern.py
pypi
from . import util from .compat import Collection, iterkeys, izip_longest, string_types, unicode class PathSpec(object): """ The :class:`PathSpec` class is a wrapper around a list of compiled :class:`.Pattern` instances. """ def __init__(self, patterns): """ Initializes the :class...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/thirdparty/pathspec/pathspec.py
0.895557
0.48871
pathspec.py
pypi
import logging from typing import Set, Callable default_logger = logging.getLogger(__name__) DEFAULT_MAX_URI_CACHE_SIZE_BYTES = (1024 ** 3) * 10 # 10 GB class URICache: """ Caches URIs up to a specified total size limit. URIs are represented by strings. Each URI has an associated size on disk. W...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/uri_cache.py
0.865764
0.256166
uri_cache.py
pypi
from abc import ABC, abstractstaticmethod from typing import Tuple from ray.util.annotations import DeveloperAPI from ray._private.runtime_env.context import RuntimeEnvContext # TODO(SongGuyang): This function exists in both C++ and Python. # We should make this logic clearly. def encode_plugin_uri(plugin: str, uri:...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/plugin.py
0.766206
0.188735
plugin.py
pypi
from enum import Enum from tempfile import TemporaryDirectory from filelock import FileLock import hashlib import logging import os from pathlib import Path import shutil from typing import Callable, List, Optional, Tuple from urllib.parse import urlparse from zipfile import ZipFile from ray.experimental.internal_kv im...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/packaging.py
0.795539
0.211519
packaging.py
pypi
import logging from pathlib import Path import sys from typing import Dict, List, Optional, Union from collections import OrderedDict import yaml logger = logging.getLogger(__name__) def validate_uri(uri: str): if not isinstance(uri, str): raise TypeError( "URIs for working_dir and py_module...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/validation.py
0.66072
0.34139
validation.py
pypi
from typing import Any, Callable, Dict, Optional, Type, Union import ray from ray.ml.config import RunConfig from ray.ml.trainer import Trainer from ray.tune import TuneError from ray.tune.result_grid import ResultGrid from ray.tune.trainable import Trainable from ray.tune.impl.tuner_internal import TunerInternal fro...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/tuner.py
0.846625
0.298325
tuner.py
pypi
import logging import uuid from functools import partial from types import FunctionType from typing import Optional import ray import ray.cloudpickle as pickle from ray.experimental.internal_kv import ( _internal_kv_initialized, _internal_kv_get, _internal_kv_put, ) from ray.tune.error import TuneError fr...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/registry.py
0.833223
0.184768
registry.py
pypi
import click import logging import operator import os import shutil import subprocess from datetime import datetime import pandas as pd from pandas.api.types import is_string_dtype, is_numeric_dtype from ray.tune.result import ( DEFAULT_EXPERIMENT_INFO_KEYS, DEFAULT_RESULT_KEYS, CONFIG_PREFIX, ) from ray.t...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/commands.py
0.581541
0.178812
commands.py
pypi
from __future__ import print_function import datetime from typing import Dict, List, Optional, Union import collections import os import sys import numpy as np import time from ray.util.annotations import PublicAPI, DeveloperAPI from ray.util.queue import Queue from ray.tune.callback import Callback from ray.tune.l...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/progress_reporter.py
0.8815
0.215536
progress_reporter.py
pypi
from abc import abstractmethod import logging from typing import Dict, List, Optional import warnings from ray.util.annotations import DeveloperAPI from ray.tune.trial import Trial, Checkpoint logger = logging.getLogger(__name__) # Signals when a class is directly inherited from TrialExecutor. # A warning is printe...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/trial_executor.py
0.917446
0.21391
trial_executor.py
pypi
from contextlib import contextmanager import inspect import os import logging import traceback from ray.util.debug import log_once from ray.util.annotations import PublicAPI, DeveloperAPI logger = logging.getLogger(__name__) _session = None @PublicAPI def is_session_enabled() -> bool: """Returns True if runnin...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/session.py
0.777173
0.164248
session.py
pypi
import heapq import gc import logging from ray.tune.utils.util import flatten_dict logger = logging.getLogger(__name__) class Checkpoint: """Describes a checkpoint of trial state. Checkpoint may be saved in different storage. Attributes: storage (str): Storage type. value (str): If sto...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/checkpoint_manager.py
0.901312
0.345105
checkpoint_manager.py
pypi
import logging from copy import copy from inspect import signature from math import isclose from typing import Any, Callable, Dict, List, Optional, Sequence, Union import numpy as np # Backwards compatibility try: # Added in numpy>=1.17 but we require numpy>=1.16 np_random_generator = np.random.Generator ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/sample.py
0.913866
0.30133
sample.py
pypi
import logging from functools import lru_cache import os import ray import time from typing import Dict from ray.tune.cluster_info import is_ray_cluster from ray.tune.trial import Trial logger = logging.getLogger(__name__) # Ideally we want to use @cache; but it's only available for python 3.9. # Caching is only he...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/insufficient_resources_manager.py
0.576304
0.179423
insufficient_resources_manager.py
pypi
from typing import Dict, Optional import time from collections import defaultdict, deque import numpy as np from ray import logger from ray.util.annotations import PublicAPI @PublicAPI class Stopper: """Base class for implementing a Tune experiment stopper. Allows users to implement experiment-level stoppin...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/stopper.py
0.980205
0.250904
stopper.py
pypi
from typing import TYPE_CHECKING, Dict, List, Optional from abc import ABC import warnings from ray.tune.checkpoint_manager import Checkpoint from ray.util.annotations import PublicAPI if TYPE_CHECKING: from ray.tune.trial import Trial from ray.tune.stopper import Stopper @PublicAPI(stability="beta") class ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/callback.py
0.926191
0.471771
callback.py
pypi
from ray.tune.error import TuneError from ray.tune.tune import run_experiments, run from ray.tune.syncer import SyncConfig from ray.tune.experiment import Experiment from ray.tune.analysis import Analysis, ExperimentAnalysis from ray.tune.stopper import Stopper from ray.tune.registry import register_env, register_train...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/__init__.py
0.641647
0.217795
__init__.py
pypi
from typing import Optional from ray.ml.checkpoint import Checkpoint from ray.ml.result import Result from ray.tune import ExperimentAnalysis from ray.tune.error import TuneError from ray.tune.trial import Trial from ray.util import PublicAPI @PublicAPI(stability="alpha") class ResultGrid: """A set of ``Result``...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/result_grid.py
0.949071
0.369486
result_grid.py
pypi
import os # fmt: off # __sphinx_doc_begin__ # (Optional/Auto-filled) training is terminated. Filled only if not provided. DONE = "done" # (Optional) Enum for user controlled checkpoint SHOULD_CHECKPOINT = "should_checkpoint" # (Auto-filled) The hostname of the machine hosting the training process. HOSTNAME = "hostna...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/result.py
0.44553
0.228845
result.py
pypi
from collections import namedtuple import logging import json from numbers import Number # For compatibility under py2 to consider unicode as str from typing import Optional from six import string_types from ray._private.resource_spec import NODE_ID_PREFIX from ray.tune import TuneError logger = logging.getLogger(_...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/resources.py
0.877759
0.22194
resources.py
pypi
import copy import logging from typing import Dict, List, Optional, Union from ray.tune.error import TuneError from ray.tune.experiment import Experiment, convert_to_experiment_list from ray.tune.config_parser import make_parser, create_trial_from_spec from ray.tune.suggest.search import SearchAlgorithm from ray.tune....
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/search_generator.py
0.940223
0.252183
search_generator.py
pypi
import copy import logging import math # use cloudpickle instead of pickle to make BOHB obj # pickleable from ray import cloudpickle from typing import Dict, List, Optional, Union from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import ( Categorical, Domain, Float, Integer, LogUnif...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/bohb.py
0.925961
0.327346
bohb.py
pypi
import logging import pickle from typing import Dict, List, Optional, Union import numpy as np import pandas as pd from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import ( Categorical, Domain, Float, Integer, LogUniform, Quantized, Uniform, ) from ray.tune.suggest.suggestio...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/hebo.py
0.935287
0.431824
hebo.py
pypi
import copy import logging from typing import Dict, List, Optional, Tuple import ray import ray.cloudpickle as pickle from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import Categorical, Domain, Float, Integer, Quantized, Uniform from ray.tune.suggest.suggestion import ( UNRESOLVED_SEARCH_SPACE, ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/zoopt.py
0.926943
0.28855
zoopt.py
pypi
from typing import Any, Dict, List, Optional import numpy as np import copy import logging from functools import partial import pickle from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import ( Categorical, Domain, Float, Integer, LogUniform, Normal, Quantized, Uniform, )...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/hyperopt.py
0.927042
0.357764
hyperopt.py
pypi
import copy import os import logging import pickle from typing import Dict, List, Optional, Union try: import sigopt as sgo Connection = sgo.Connection except ImportError: sgo = None Connection = None from ray.tune.result import DEFAULT_METRIC from ray.tune.suggest import Searcher logger = logging.g...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/sigopt.py
0.909506
0.291999
sigopt.py
pypi
from collections import defaultdict import logging import pickle import json from typing import Dict, List, Optional, Tuple, Any from ray.tune import ExperimentAnalysis from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import Domain, Float, Quantized from ray.tune.suggest.suggestion import ( UNRESOLV...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/bayesopt.py
0.937354
0.331309
bayesopt.py
pypi
import copy import pickle from typing import Dict, List, Optional, Union from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import Categorical, Float, Integer, LogUniform, Quantized, Uniform from ray.tune.suggest.suggestion import ( UNRESOLVED_SEARCH_SPACE, UNDEFINED_METRIC_MODE, UNDEFINED_SEA...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/ax.py
0.921596
0.388212
ax.py
pypi
import copy import logging import numpy as np import pickle from typing import Dict, List, Optional, Tuple, Union, Any from ray.tune.result import DEFAULT_METRIC from ray.tune.sample import Categorical, Domain, Float, Integer, Quantized, LogUniform from ray.tune.suggest import Searcher from ray.tune.suggest.suggestion...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/skopt.py
0.916946
0.427456
skopt.py
pypi
from ray._private.utils import get_function_args from ray.tune.suggest.search import SearchAlgorithm from ray.tune.suggest.basic_variant import BasicVariantGenerator from ray.tune.suggest.suggestion import Searcher, ConcurrencyLimiter from ray.tune.suggest.search_generator import SearchGenerator from ray.tune.suggest.v...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/__init__.py
0.882288
0.241098
__init__.py
pypi
import copy import glob import itertools import os import uuid from typing import Dict, List, Optional, Union import warnings import numpy as np from ray.tune.error import TuneError from ray.tune.experiment import Experiment, convert_to_experiment_list from ray.tune.config_parser import make_parser, create_trial_from_...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/basic_variant.py
0.879645
0.242295
basic_variant.py
pypi
import copy import logging from typing import Dict, List, Optional import numpy as np from ray.tune.suggest.suggestion import Searcher from ray.tune.suggest.util import set_search_properties_backwards_compatible logger = logging.getLogger(__name__) TRIAL_INDEX = "__trial_index__" """str: A constant value representi...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/repeater.py
0.93651
0.380241
repeater.py
pypi
from collections import Counter from typing import Dict, List, Union from tensorflow.keras.callbacks import Callback from ray import tune import os class TuneCallback(Callback): """Base class for Tune's Keras callbacks.""" _allowed = [ "batch_begin", "batch_end", "epoch_begin", ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/keras.py
0.952937
0.207536
keras.py
pypi
from typing import Dict, Callable, Optional import logging import ray from ray.tune.trainable import Trainable from ray.tune.logger import Logger, LoggerCallback from ray.tune.result import TRAINING_ITERATION, TIMESTEPS_TOTAL from ray.tune.trial import Trial from ray.util.annotations import Deprecated from ray.util.ml...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/mlflow.py
0.931793
0.397588
mlflow.py
pypi
import os import pickle from collections.abc import Sequence from multiprocessing import Process, Queue from numbers import Number from typing import Any, Callable, Dict, List, Optional, Tuple import numpy as np import urllib from ray import logger from ray.tune import Trainable from ray.tune.function_runner import Fu...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/wandb.py
0.81119
0.287755
wandb.py
pypi
import logging from typing import Dict, List, Optional, Union from pytorch_lightning import Callback, Trainer, LightningModule from ray import tune import os logger = logging.getLogger(__name__) class TuneCallback(Callback): """Base class for Tune's PyTorch Lightning callbacks.""" _allowed = [ "in...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/pytorch_lightning.py
0.923338
0.305166
pytorch_lightning.py
pypi
from typing import Dict, List, Union, Callable, Optional from ray import tune import os from ray.tune.utils import flatten_dict from lightgbm.callback import CallbackEnv from lightgbm.basic import Booster class TuneCallback: """Base class for Tune's LightGBM callbacks.""" pass class TuneReportCallback(Tu...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/lightgbm.py
0.966291
0.270827
lightgbm.py
pypi
import os from typing import Any, Optional, Tuple, List import subprocess from ray import logger from ray.autoscaler._private.command_runner import KubernetesCommandRunner from ray.tune.syncer import NodeSyncer from ray.tune.sync_client import SyncClient from ray.util import get_node_ip_address def try_import_kubern...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/kubernetes.py
0.884214
0.183319
kubernetes.py
pypi
from typing import Callable, Dict, List, Union, Optional from collections import OrderedDict from ray import tune import os from ray.tune.utils import flatten_dict from xgboost.core import Booster try: from xgboost.callback import TrainingCallback except ImportError: class TrainingCallback: pass c...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/xgboost.py
0.97377
0.304688
xgboost.py
pypi
from contextlib import contextmanager import os import logging import shutil import tempfile from typing import Callable, Dict, Generator, Optional, Type import torch from datetime import timedelta import ray from ray import tune from ray.tune.result import RESULT_DUPLICATE from ray.tune.logger import NoopLogger from...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/torch.py
0.845751
0.238694
torch.py
pypi
from typing import Callable, Dict, Type, Optional from contextlib import contextmanager import os import logging import shutil import tempfile from filelock import FileLock import ray from ray import tune from ray.tune.function_runner import wrap_function from ray.tune.logger import NoopLogger from ray.tune.result i...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/horovod.py
0.891628
0.250511
horovod.py
pypi
from typing import Dict, List, Union from ray import tune import mxnet from mxnet.model import save_checkpoint, BatchEndParam import numpy as np import os class TuneCallback: """Base class for Tune's MXNet callbacks.""" pass class TuneReportCallback(TuneCallback): """MXNet to Ray Tune reporting call...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/mxnet.py
0.955413
0.288929
mxnet.py
pypi
import json import logging import ray import os from ray.util.ml_utils.util import find_free_port from ray.tune.function_runner import wrap_function from ray.tune.resources import Resources from ray.tune.result import RESULT_DUPLICATE from ray.tune.trainable import DistributedTrainable from ray.tune.utils import detec...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/tensorflow.py
0.749729
0.327507
tensorflow.py
pypi
import copy import os from typing import Any, Callable, Dict, Optional, Type, Union import ray.cloudpickle as pickle from ray.ml.config import RunConfig from ray.ml.trainer import Trainer from ray.tune import Experiment, TuneError, ExperimentAnalysis from ray.tune.impl.utils import execute_dataset from ray.tune.result...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/impl/tuner_internal.py
0.900351
0.216094
tuner_internal.py
pypi
from django.db import models class JobRecord(models.Model): """Information of an AutoML Job.""" job_id = models.CharField(max_length=50) name = models.CharField(max_length=20) user = models.CharField(max_length=20) type = models.CharField(max_length=20) start_time = models.CharField(max_lengt...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/models/models.py
0.828523
0.231213
models.py
pypi
from django.shortcuts import HttpResponse from ray.tune.automlboard.models.models import JobRecord, TrialRecord from ray.tune.trial import Trial import json def query_job(request): """Rest API to query the job info, with the given job_id. The url pattern should be like this: curl http://<server>:<port...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/frontend/query.py
0.48121
0.322259
query.py
pypi
import logging import os import time from threading import Thread from ray.tune.automlboard.common.exception import CollectorError from ray.tune.automlboard.common.utils import ( parse_json, parse_multiple_json, timestamp2date, ) from ray.tune.automlboard.models.models import JobRecord, TrialRecord, Resul...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/backend/collector.py
0.744563
0.158435
collector.py
pypi
import logging import json import os import time def dump_json(json_info, json_file, overwrite=True): """Dump a whole json record into the given file. Overwrite the file if the overwrite flag set. Args: json_info (dict): Information dict to be dumped. json_file (str): File path to be dum...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/common/utils.py
0.563138
0.185007
utils.py
pypi
import json import logging import os import warnings import traceback from numbers import Number from typing import Any, Dict, List, Optional, Tuple from ray.ml.checkpoint import Checkpoint from ray.tune.cloud import TrialCheckpoint from ray.util.debug import log_once from ray.tune.syncer import SyncConfig from ray.tu...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/analysis/experiment_analysis.py
0.892475
0.182098
experiment_analysis.py
pypi
# __import_lightning_begin__ import math import torch import pytorch_lightning as pl from filelock import FileLock from torch.utils.data import DataLoader, random_split from torch.nn import functional as F from torchvision.datasets import MNIST from torchvision import transforms import os # __import_lightning_end__ ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_pytorch_lightning.py
0.880271
0.439687
mnist_pytorch_lightning.py
pypi
import mxnet as mx from ray import tune, logger from ray.tune.integration.mxnet import TuneCheckpointCallback, TuneReportCallback from ray.tune.schedulers import ASHAScheduler def train_mnist_mxnet(config, mnist, num_epochs=10): batch_size = config["batch_size"] train_iter = mx.io.NDArrayIter( mnist[...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mxnet_example.py
0.730386
0.406332
mxnet_example.py
pypi
import argparse import tensorflow as tf import numpy as np import ray from ray import tune from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.integration.keras import TuneReportCheckpointCallback from ray.tune.integration.tensorflow import DistributedTrainableCreator, get_num_workers def mnist_data...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tf_distributed_keras_example.py
0.848831
0.460774
tf_distributed_keras_example.py
pypi
import argparse import os from filelock import FileLock from tensorflow.keras.layers import Dense, Flatten, Conv2D from tensorflow.keras import Model from tensorflow.keras.datasets.mnist import load_data from ray import tune MAX_TRAIN_BATCH = 10 class MyModel(Model): def __init__(self, hiddens=128): s...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tf_mnist_example.py
0.903664
0.407098
tf_mnist_example.py
pypi
import argparse import json import time import os import numpy as np import ray from ray import tune from ray.tune import Trainable from ray.tune.schedulers.hb_bohb import HyperBandForBOHB from ray.tune.suggest.bohb import TuneBOHB class MyTrainableClass(Trainable): """Example agent whose learning curve is a r...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/bohb_example.py
0.850531
0.307267
bohb_example.py
pypi
import torch import torch.nn as nn import os import numpy as np import torchvision from torch.utils.data import DataLoader import torchvision.transforms as transforms import ray from ray import tune from ray.tune.schedulers import create_scheduler from ray.tune.integration.horovod import ( DistributedTrainableCre...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/horovod_cifar_pbt_example.py
0.757615
0.418875
horovod_cifar_pbt_example.py
pypi
import time import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.optuna import OptunaSearch def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objectiv...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_example.py
0.753194
0.273485
optuna_example.py
pypi
import math import torch from filelock import FileLock from torch.nn import functional as F from torchmetrics import Accuracy import pytorch_lightning as pl from pl_bolts.datamodules.mnist_datamodule import MNISTDataModule import os from ray.tune.integration.pytorch_lightning import TuneReportCallback from ray import...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_ptl_mini.py
0.902742
0.47859
mnist_ptl_mini.py
pypi
from __future__ import print_function import argparse import os import torch import torch.optim as optim import ray from ray import tune from ray.tune.schedulers import ASHAScheduler from ray.tune.examples.mnist_pytorch import train, test, get_data_loaders, ConvNet # Change these values if you want the training to r...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_pytorch_trainable.py
0.852935
0.278272
mnist_pytorch_trainable.py
pypi
import time from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.skopt import SkOptSearch def evaluation_fn(step, width, height): time.sleep(0.1) return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_o...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/skopt_example.py
0.811713
0.287565
skopt_example.py
pypi
import time import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.flaml import BlendSearch def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objective(...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/blendsearch_example.py
0.74008
0.317942
blendsearch_example.py
pypi
# __tutorial_imports_begin__ import argparse import os import numpy as np import torch import torch.optim as optim from ray.tune.examples.mnist_pytorch import train, test, ConvNet, get_data_loaders from ray import tune from ray.tune.schedulers import PopulationBasedTraining from ray.tune.trial import ExportFormat # ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_convnet_function_example.py
0.870184
0.38659
pbt_convnet_function_example.py
pypi
from __future__ import print_function from tensorflow.keras.models import Sequential, Model, load_model from tensorflow.keras.layers import Embedding from tensorflow.keras.layers import Input, Activation, Dense, Permute, Dropout from tensorflow.keras.layers import add, dot, concatenate from tensorflow.keras.layers imp...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_memnn_example.py
0.715325
0.366391
pbt_memnn_example.py
pypi
from typing import Dict, List import sklearn.datasets import sklearn.metrics import os import numpy as np from ray.tune.schedulers import ASHAScheduler from sklearn.model_selection import train_test_split import xgboost as xgb from ray import tune from ray.tune.integration.xgboost import ( TuneReportCheckpointCall...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/xgboost_example.py
0.903018
0.550003
xgboost_example.py
pypi
import random from ray import tune from ray.tune.schedulers import PopulationBasedTraining if __name__ == "__main__": # Postprocess the perturbed config to ensure it's still valid def explore(config): # ensure we collect enough timesteps to do sgd if config["train_batch_size"] < config["sgd_m...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_ppo_example.py
0.675015
0.261985
pbt_ppo_example.py
pypi
import argparse import time import ray from ray import tune from ray.tune.schedulers import AsyncHyperBandScheduler def evaluation_fn(step, width, height): time.sleep(0.1) return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objective(config): # Hyperparameters width, height = config[...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/async_hyperband_example.py
0.731826
0.268216
async_hyperband_example.py
pypi
import time import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.flaml import CFO def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objective(config):...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/cfo_example.py
0.752013
0.28208
cfo_example.py
pypi
import sys import time from ray import tune from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.sigopt import SigOptSearch def evaluate(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.01 def easy_objective(config): # Hyperparameters width, height =...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/sigopt_example.py
0.583085
0.261213
sigopt_example.py
pypi
"""Examples using MLfowLoggerCallback and mlflow_mixin. """ import os import tempfile import time import mlflow from ray import tune from ray.tune.integration.mlflow import MLflowLoggerCallback, mlflow_mixin def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def e...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mlflow_example.py
0.729905
0.471467
mlflow_example.py
pypi
import os import time import json import argparse from ray import tune def evaluation_fn(step, width, height): time.sleep(0.1) return (0.1 + width * step / 100) ** (-1) + height * 0.1 def train_func(config, checkpoint_dir=None): start = 0 width, height = config["width"], config["height"] if ch...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/custom_func_checkpointing.py
0.669205
0.174621
custom_func_checkpointing.py
pypi
import argparse import numpy as np import time import logging import os import ray from ray import tune from ray.tune import Trainable from ray.tune.sync_client import get_sync_client from ray import cloudpickle logger = logging.getLogger(__name__) class MockDurableTrainable(Trainable): """Mocks the storage cli...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/durable_trainable_example.py
0.753648
0.218732
durable_trainable_example.py
pypi
# __import_begin__ from functools import partial import numpy as np import os import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim from filelock import FileLock from torch.utils.data import random_split import torchvision import torchvision.transforms as transforms import ray ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/cifar10_pytorch.py
0.888819
0.327897
cifar10_pytorch.py
pypi
import argparse import json import os import numpy as np import ray from ray import tune from ray.tune.schedulers import HyperBandScheduler def train(config, checkpoint_dir=None): step = 0 if checkpoint_dir: with open(os.path.join(checkpoint_dir, "checkpoint")) as f: step = json.loads(f...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/hyperband_function_example.py
0.656988
0.224523
hyperband_function_example.py
pypi
import numpy as np import time from ray import tune from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.ax import AxSearch def hartmann6(x): alpha = np.array([1.0, 1.2, 3.0, 3.2]) A = np.array( [ [10, 3, 17, 3.5, 1.7, 8], [0.05, 10, 17, 0.1, 8, 14], ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/ax_example.py
0.646795
0.393356
ax_example.py
pypi
import argparse import os from filelock import FileLock from tensorflow.keras.datasets import mnist import ray from ray import tune from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.integration.keras import TuneReportCallback def train_mnist(config): # https://github.com/tensorflow/tensorflo...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tune_mnist_keras.py
0.812459
0.379206
tune_mnist_keras.py
pypi
# flake8: noqa # fmt: off # __tutorial_imports_begin__ import argparse import os import numpy as np import torch import torch.optim as optim from torchvision import datasets from ray.tune.examples.mnist_pytorch import train, test, ConvNet,\ get_data_loaders import ray from ray import tune from ray.tune.scheduler...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_convnet_example.py
0.830285
0.464841
pbt_convnet_example.py
pypi
import time from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.nevergrad import NevergradSearch def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objective(con...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/nevergrad_example.py
0.791821
0.274572
nevergrad_example.py
pypi
import argparse import logging import os import torch import torch.optim as optim from torch.nn.parallel import DistributedDataParallel import ray from ray import tune from ray.tune.examples.mnist_pytorch import train, test, get_data_loaders, ConvNet from ray.tune.integration.torch import ( DistributedTrainableCre...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/ddp_mnist_torch.py
0.734024
0.310459
ddp_mnist_torch.py
pypi
HPO, and MLflow autologging all together.""" import os import tempfile import pytorch_lightning as pl from pl_bolts.datamodules import MNISTDataModule import mlflow from ray import tune from ray.tune.integration.mlflow import mlflow_mixin from ray.tune.integration.pytorch_lightning import TuneReportCallback from ray...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mlflow_ptl.py
0.687525
0.535645
mlflow_ptl.py
pypi
import time import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.hyperopt import HyperOptSearch from hyperopt import hp def f_unpack_dict(dct): """ Unpacks all sub-dictionaries in given dictionary recursi...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/hyperopt_conditional_search_space_example.py
0.905034
0.312003
hyperopt_conditional_search_space_example.py
pypi
import time from typing import Dict, Optional, Any import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.optuna import OptunaSearch def evaluation_fn(step, width, height, mult=1): return (0.1 + width * step / ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_define_by_run_example.py
0.920598
0.350616
optuna_define_by_run_example.py
pypi
import argparse import tempfile from unittest.mock import MagicMock import numpy as np import wandb from ray import tune from ray.tune import Trainable from ray.tune.integration.wandb import ( WandbLoggerCallback, WandbTrainableMixin, wandb_mixin, ) def train_function(config, checkpoint_dir=None): f...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/wandb_example.py
0.771628
0.409693
wandb_example.py
pypi
from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import time from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.schedulers import AsyncHyperBandScheduler from ray.tune.suggest.dragonfly import DragonflySearc...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/dragonfly_example.py
0.729134
0.19095
dragonfly_example.py
pypi
from __future__ import print_function import argparse import random import mxnet as mx import numpy as np from mxnet import gluon, init from mxnet import autograd as ag from mxnet.gluon import nn from mxnet.gluon.data.vision import transforms from gluoncv.model_zoo import get_model from gluoncv.data import transform...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tune_cifar10_gluon.py
0.873768
0.264142
tune_cifar10_gluon.py
pypi
import time import ray from ray import tune from ray.tune.suggest import ConcurrencyLimiter from ray.tune.suggest.optuna import OptunaSearch def evaluation_fn(step, width, height): return (0.1 + width * step / 100) ** (-1) + height * 0.1 def easy_objective(config): # Hyperparameters width, height = con...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_multiobjective_example.py
0.732783
0.272464
optuna_multiobjective_example.py
pypi
from __future__ import print_function import argparse import numpy as np import tensorflow as tf from tensorflow.keras.datasets import cifar10 from tensorflow.keras.layers import Input, Dense, Dropout, Flatten from tensorflow.keras.layers import Convolution2D, MaxPooling2D from tensorflow.keras.models import Model, l...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_tune_cifar10_with_keras.py
0.9151
0.443721
pbt_tune_cifar10_with_keras.py
pypi
import os import ray from ray import tune from ray.tune import CLIReporter from ray.tune.examples.pbt_transformers.utils import ( download_data, build_compute_metrics_fn, ) from ray.tune.schedulers import PopulationBasedTraining from transformers import ( glue_tasks_num_labels, AutoConfig, AutoMode...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_transformers/pbt_transformers.py
0.578805
0.258431
pbt_transformers.py
pypi
from typing import Dict, Optional from copy import deepcopy import logging import numpy as np import pandas as pd from ray.tune import TuneError from ray.tune.schedulers import PopulationBasedTraining def import_pb2_dependencies(): try: import GPy except ImportError: GPy = None try: ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/pb2.py
0.849909
0.58255
pb2.py
pypi
import collections import logging from typing import Dict, List, Optional import numpy as np from ray.tune import trial_runner from ray.tune.result import DEFAULT_METRIC from ray.tune.trial import Trial from ray.tune.schedulers.trial_scheduler import FIFOScheduler, TrialScheduler logger = logging.getLogger(__name__)...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/median_stopping_rule.py
0.943073
0.345202
median_stopping_rule.py
pypi
import numpy as np from scipy.optimize import minimize import GPy from GPy.kern import Kern from GPy.core import Param from sklearn.metrics import pairwise_distances from sklearn.metrics.pairwise import euclidean_distances class TV_SquaredExp(Kern): """Time varying squared exponential kernel. For more info s...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/pb2_utils.py
0.868715
0.614076
pb2_utils.py
pypi
from typing import Dict, Optional from ray.tune import trial_runner from ray.tune.result import DEFAULT_METRIC from ray.tune.trial import Trial class TrialScheduler: """Interface for implementing a Trial Scheduler class.""" CONTINUE = "CONTINUE" #: Status for continuing trial execution PAUSE = "PAUSE" ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/trial_scheduler.py
0.883645
0.24164
trial_scheduler.py
pypi
from ray._private.utils import get_function_args from ray.tune.schedulers.trial_scheduler import TrialScheduler, FIFOScheduler from ray.tune.schedulers.hyperband import HyperBandScheduler from ray.tune.schedulers.hb_bohb import HyperBandForBOHB from ray.tune.schedulers.async_hyperband import AsyncHyperBandScheduler, AS...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/__init__.py
0.905387
0.440469
__init__.py
pypi
import time import copy import logging from ray.tune.trial import Trial from ray.tune.suggest import SearchAlgorithm from ray.tune.experiment import convert_to_experiment_list from ray.tune.suggest.variant_generator import generate_variants from ray.tune.config_parser import make_parser, create_trial_from_spec logger...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automl/search_policy.py
0.713531
0.192748
search_policy.py
pypi
import random import logging import numpy as np from ray.tune import grid_search logger = logging.getLogger(__name__) class ParameterSpace: """Base class of a single parameter's search space.""" def __init__(self, name): """Initialize ParameterSpace. Arguments: name (str): Name...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automl/search_space.py
0.787155
0.457076
search_space.py
pypi
from typing import Dict, List, Union import copy import glob import logging import os import inspect import threading import time import uuid from collections import defaultdict from datetime import datetime from threading import Thread from typing import Optional import numpy as np import ray import psutil from ray....
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/utils/util.py
0.813535
0.154695
util.py
pypi
from typing import Optional, Dict, Any import logging import os import time import ray from ray import ray_constants from ray._private.resource_spec import NODE_ID_PREFIX from ray.tune.resources import Resources logger = logging.getLogger(__name__) TUNE_STATE_REFRESH_PERIOD = 10 # Refresh resources every 10 s def...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/utils/resource_updater.py
0.888457
0.170128
resource_updater.py
pypi
from typing import Any, List, Tuple, Dict, Optional class CommandRunnerInterface: """Interface to run commands on a remote cluster node. **Important**: This is an INTERNAL API that is only exposed for the purpose of implementing custom node providers. It is not allowed to call into CommandRunner meth...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/command_runner.py
0.916717
0.310707
command_runner.py
pypi
import logging from types import ModuleType from typing import Any, Dict, List, Optional from ray.autoscaler.command_runner import CommandRunnerInterface from ray.autoscaler._private.command_runner import SSHCommandRunner, DockerCommandRunner logger = logging.getLogger(__name__) class NodeProvider: """Interface...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/node_provider.py
0.941654
0.339828
node_provider.py
pypi
from enum import Enum, auto from typing import Any, Callable, Dict, List, Optional, Union from ray.autoscaler._private.cli_logger import cli_logger class CreateClusterEvent(Enum): """Events to track in ray.autoscaler.sdk.create_or_update_cluster. Attributes: up_started : Invoked at the beginning of ...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/_private/event_system.py
0.909959
0.219578
event_system.py
pypi
from ray.autoscaler._private import constants from typing import List, Set, Tuple class NodeTracker: """Map nodes to their corresponding logs. We need to be a little careful here. At an given point in time, node_id <-> ip can be interchangeably used, but the node_id -> ip relation is not bijective _a...
/ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/_private/node_tracker.py
0.911031
0.482673
node_tracker.py
pypi