code stringlengths 114 1.05M | path stringlengths 3 312 | quality_prob float64 0.5 0.99 | learning_prob float64 0.2 1 | filename stringlengths 3 168 | kind stringclasses 1
value |
|---|---|---|---|---|---|
import re
from .compat import unicode
class Pattern(object):
"""
The :class:`Pattern` class is the abstract definition of a pattern.
"""
# Make the class dict-less.
__slots__ = ("include",)
def __init__(self, include):
"""
Initializes the :class:`Pattern` instance.
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/thirdparty/pathspec/pattern.py | 0.894138 | 0.489626 | pattern.py | pypi |
from . import util
from .compat import Collection, iterkeys, izip_longest, string_types, unicode
class PathSpec(object):
"""
The :class:`PathSpec` class is a wrapper around a list of compiled
:class:`.Pattern` instances.
"""
def __init__(self, patterns):
"""
Initializes the :class... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/thirdparty/pathspec/pathspec.py | 0.895557 | 0.48871 | pathspec.py | pypi |
import logging
from typing import Set, Callable
default_logger = logging.getLogger(__name__)
DEFAULT_MAX_URI_CACHE_SIZE_BYTES = (1024 ** 3) * 10 # 10 GB
class URICache:
"""
Caches URIs up to a specified total size limit.
URIs are represented by strings. Each URI has an associated size on disk.
W... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/uri_cache.py | 0.865764 | 0.256166 | uri_cache.py | pypi |
from abc import ABC, abstractstaticmethod
from typing import Tuple
from ray.util.annotations import DeveloperAPI
from ray._private.runtime_env.context import RuntimeEnvContext
# TODO(SongGuyang): This function exists in both C++ and Python.
# We should make this logic clearly.
def encode_plugin_uri(plugin: str, uri:... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/plugin.py | 0.766206 | 0.188735 | plugin.py | pypi |
from enum import Enum
from tempfile import TemporaryDirectory
from filelock import FileLock
import hashlib
import logging
import os
from pathlib import Path
import shutil
from typing import Callable, List, Optional, Tuple
from urllib.parse import urlparse
from zipfile import ZipFile
from ray.experimental.internal_kv im... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/packaging.py | 0.795539 | 0.211519 | packaging.py | pypi |
import logging
from pathlib import Path
import sys
from typing import Dict, List, Optional, Union
from collections import OrderedDict
import yaml
logger = logging.getLogger(__name__)
def validate_uri(uri: str):
if not isinstance(uri, str):
raise TypeError(
"URIs for working_dir and py_module... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/_private/runtime_env/validation.py | 0.66072 | 0.34139 | validation.py | pypi |
from typing import Any, Callable, Dict, Optional, Type, Union
import ray
from ray.ml.config import RunConfig
from ray.ml.trainer import Trainer
from ray.tune import TuneError
from ray.tune.result_grid import ResultGrid
from ray.tune.trainable import Trainable
from ray.tune.impl.tuner_internal import TunerInternal
fro... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/tuner.py | 0.846625 | 0.298325 | tuner.py | pypi |
import logging
import uuid
from functools import partial
from types import FunctionType
from typing import Optional
import ray
import ray.cloudpickle as pickle
from ray.experimental.internal_kv import (
_internal_kv_initialized,
_internal_kv_get,
_internal_kv_put,
)
from ray.tune.error import TuneError
fr... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/registry.py | 0.833223 | 0.184768 | registry.py | pypi |
import click
import logging
import operator
import os
import shutil
import subprocess
from datetime import datetime
import pandas as pd
from pandas.api.types import is_string_dtype, is_numeric_dtype
from ray.tune.result import (
DEFAULT_EXPERIMENT_INFO_KEYS,
DEFAULT_RESULT_KEYS,
CONFIG_PREFIX,
)
from ray.t... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/commands.py | 0.581541 | 0.178812 | commands.py | pypi |
from __future__ import print_function
import datetime
from typing import Dict, List, Optional, Union
import collections
import os
import sys
import numpy as np
import time
from ray.util.annotations import PublicAPI, DeveloperAPI
from ray.util.queue import Queue
from ray.tune.callback import Callback
from ray.tune.l... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/progress_reporter.py | 0.8815 | 0.215536 | progress_reporter.py | pypi |
from abc import abstractmethod
import logging
from typing import Dict, List, Optional
import warnings
from ray.util.annotations import DeveloperAPI
from ray.tune.trial import Trial, Checkpoint
logger = logging.getLogger(__name__)
# Signals when a class is directly inherited from TrialExecutor.
# A warning is printe... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/trial_executor.py | 0.917446 | 0.21391 | trial_executor.py | pypi |
from contextlib import contextmanager
import inspect
import os
import logging
import traceback
from ray.util.debug import log_once
from ray.util.annotations import PublicAPI, DeveloperAPI
logger = logging.getLogger(__name__)
_session = None
@PublicAPI
def is_session_enabled() -> bool:
"""Returns True if runnin... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/session.py | 0.777173 | 0.164248 | session.py | pypi |
import heapq
import gc
import logging
from ray.tune.utils.util import flatten_dict
logger = logging.getLogger(__name__)
class Checkpoint:
"""Describes a checkpoint of trial state.
Checkpoint may be saved in different storage.
Attributes:
storage (str): Storage type.
value (str): If sto... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/checkpoint_manager.py | 0.901312 | 0.345105 | checkpoint_manager.py | pypi |
import logging
from copy import copy
from inspect import signature
from math import isclose
from typing import Any, Callable, Dict, List, Optional, Sequence, Union
import numpy as np
# Backwards compatibility
try:
# Added in numpy>=1.17 but we require numpy>=1.16
np_random_generator = np.random.Generator
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/sample.py | 0.913866 | 0.30133 | sample.py | pypi |
import logging
from functools import lru_cache
import os
import ray
import time
from typing import Dict
from ray.tune.cluster_info import is_ray_cluster
from ray.tune.trial import Trial
logger = logging.getLogger(__name__)
# Ideally we want to use @cache; but it's only available for python 3.9.
# Caching is only he... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/insufficient_resources_manager.py | 0.576304 | 0.179423 | insufficient_resources_manager.py | pypi |
from typing import Dict, Optional
import time
from collections import defaultdict, deque
import numpy as np
from ray import logger
from ray.util.annotations import PublicAPI
@PublicAPI
class Stopper:
"""Base class for implementing a Tune experiment stopper.
Allows users to implement experiment-level stoppin... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/stopper.py | 0.980205 | 0.250904 | stopper.py | pypi |
from typing import TYPE_CHECKING, Dict, List, Optional
from abc import ABC
import warnings
from ray.tune.checkpoint_manager import Checkpoint
from ray.util.annotations import PublicAPI
if TYPE_CHECKING:
from ray.tune.trial import Trial
from ray.tune.stopper import Stopper
@PublicAPI(stability="beta")
class ... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/callback.py | 0.926191 | 0.471771 | callback.py | pypi |
from ray.tune.error import TuneError
from ray.tune.tune import run_experiments, run
from ray.tune.syncer import SyncConfig
from ray.tune.experiment import Experiment
from ray.tune.analysis import Analysis, ExperimentAnalysis
from ray.tune.stopper import Stopper
from ray.tune.registry import register_env, register_train... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/__init__.py | 0.641647 | 0.217795 | __init__.py | pypi |
from typing import Optional
from ray.ml.checkpoint import Checkpoint
from ray.ml.result import Result
from ray.tune import ExperimentAnalysis
from ray.tune.error import TuneError
from ray.tune.trial import Trial
from ray.util import PublicAPI
@PublicAPI(stability="alpha")
class ResultGrid:
"""A set of ``Result``... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/result_grid.py | 0.949071 | 0.369486 | result_grid.py | pypi |
import os
# fmt: off
# __sphinx_doc_begin__
# (Optional/Auto-filled) training is terminated. Filled only if not provided.
DONE = "done"
# (Optional) Enum for user controlled checkpoint
SHOULD_CHECKPOINT = "should_checkpoint"
# (Auto-filled) The hostname of the machine hosting the training process.
HOSTNAME = "hostna... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/result.py | 0.44553 | 0.228845 | result.py | pypi |
from collections import namedtuple
import logging
import json
from numbers import Number
# For compatibility under py2 to consider unicode as str
from typing import Optional
from six import string_types
from ray._private.resource_spec import NODE_ID_PREFIX
from ray.tune import TuneError
logger = logging.getLogger(_... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/resources.py | 0.877759 | 0.22194 | resources.py | pypi |
import copy
import logging
from typing import Dict, List, Optional, Union
from ray.tune.error import TuneError
from ray.tune.experiment import Experiment, convert_to_experiment_list
from ray.tune.config_parser import make_parser, create_trial_from_spec
from ray.tune.suggest.search import SearchAlgorithm
from ray.tune.... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/search_generator.py | 0.940223 | 0.252183 | search_generator.py | pypi |
import copy
import logging
import math
# use cloudpickle instead of pickle to make BOHB obj
# pickleable
from ray import cloudpickle
from typing import Dict, List, Optional, Union
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import (
Categorical,
Domain,
Float,
Integer,
LogUnif... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/bohb.py | 0.925961 | 0.327346 | bohb.py | pypi |
import logging
import pickle
from typing import Dict, List, Optional, Union
import numpy as np
import pandas as pd
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import (
Categorical,
Domain,
Float,
Integer,
LogUniform,
Quantized,
Uniform,
)
from ray.tune.suggest.suggestio... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/hebo.py | 0.935287 | 0.431824 | hebo.py | pypi |
import copy
import logging
from typing import Dict, List, Optional, Tuple
import ray
import ray.cloudpickle as pickle
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import Categorical, Domain, Float, Integer, Quantized, Uniform
from ray.tune.suggest.suggestion import (
UNRESOLVED_SEARCH_SPACE,
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/zoopt.py | 0.926943 | 0.28855 | zoopt.py | pypi |
from typing import Any, Dict, List, Optional
import numpy as np
import copy
import logging
from functools import partial
import pickle
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import (
Categorical,
Domain,
Float,
Integer,
LogUniform,
Normal,
Quantized,
Uniform,
)... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/hyperopt.py | 0.927042 | 0.357764 | hyperopt.py | pypi |
import copy
import os
import logging
import pickle
from typing import Dict, List, Optional, Union
try:
import sigopt as sgo
Connection = sgo.Connection
except ImportError:
sgo = None
Connection = None
from ray.tune.result import DEFAULT_METRIC
from ray.tune.suggest import Searcher
logger = logging.g... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/sigopt.py | 0.909506 | 0.291999 | sigopt.py | pypi |
from collections import defaultdict
import logging
import pickle
import json
from typing import Dict, List, Optional, Tuple, Any
from ray.tune import ExperimentAnalysis
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import Domain, Float, Quantized
from ray.tune.suggest.suggestion import (
UNRESOLV... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/bayesopt.py | 0.937354 | 0.331309 | bayesopt.py | pypi |
import copy
import pickle
from typing import Dict, List, Optional, Union
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import Categorical, Float, Integer, LogUniform, Quantized, Uniform
from ray.tune.suggest.suggestion import (
UNRESOLVED_SEARCH_SPACE,
UNDEFINED_METRIC_MODE,
UNDEFINED_SEA... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/ax.py | 0.921596 | 0.388212 | ax.py | pypi |
import copy
import logging
import numpy as np
import pickle
from typing import Dict, List, Optional, Tuple, Union, Any
from ray.tune.result import DEFAULT_METRIC
from ray.tune.sample import Categorical, Domain, Float, Integer, Quantized, LogUniform
from ray.tune.suggest import Searcher
from ray.tune.suggest.suggestion... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/skopt.py | 0.916946 | 0.427456 | skopt.py | pypi |
from ray._private.utils import get_function_args
from ray.tune.suggest.search import SearchAlgorithm
from ray.tune.suggest.basic_variant import BasicVariantGenerator
from ray.tune.suggest.suggestion import Searcher, ConcurrencyLimiter
from ray.tune.suggest.search_generator import SearchGenerator
from ray.tune.suggest.v... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/__init__.py | 0.882288 | 0.241098 | __init__.py | pypi |
import copy
import glob
import itertools
import os
import uuid
from typing import Dict, List, Optional, Union
import warnings
import numpy as np
from ray.tune.error import TuneError
from ray.tune.experiment import Experiment, convert_to_experiment_list
from ray.tune.config_parser import make_parser, create_trial_from_... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/basic_variant.py | 0.879645 | 0.242295 | basic_variant.py | pypi |
import copy
import logging
from typing import Dict, List, Optional
import numpy as np
from ray.tune.suggest.suggestion import Searcher
from ray.tune.suggest.util import set_search_properties_backwards_compatible
logger = logging.getLogger(__name__)
TRIAL_INDEX = "__trial_index__"
"""str: A constant value representi... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/suggest/repeater.py | 0.93651 | 0.380241 | repeater.py | pypi |
from collections import Counter
from typing import Dict, List, Union
from tensorflow.keras.callbacks import Callback
from ray import tune
import os
class TuneCallback(Callback):
"""Base class for Tune's Keras callbacks."""
_allowed = [
"batch_begin",
"batch_end",
"epoch_begin",
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/keras.py | 0.952937 | 0.207536 | keras.py | pypi |
from typing import Dict, Callable, Optional
import logging
import ray
from ray.tune.trainable import Trainable
from ray.tune.logger import Logger, LoggerCallback
from ray.tune.result import TRAINING_ITERATION, TIMESTEPS_TOTAL
from ray.tune.trial import Trial
from ray.util.annotations import Deprecated
from ray.util.ml... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/mlflow.py | 0.931793 | 0.397588 | mlflow.py | pypi |
import os
import pickle
from collections.abc import Sequence
from multiprocessing import Process, Queue
from numbers import Number
from typing import Any, Callable, Dict, List, Optional, Tuple
import numpy as np
import urllib
from ray import logger
from ray.tune import Trainable
from ray.tune.function_runner import Fu... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/wandb.py | 0.81119 | 0.287755 | wandb.py | pypi |
import logging
from typing import Dict, List, Optional, Union
from pytorch_lightning import Callback, Trainer, LightningModule
from ray import tune
import os
logger = logging.getLogger(__name__)
class TuneCallback(Callback):
"""Base class for Tune's PyTorch Lightning callbacks."""
_allowed = [
"in... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/pytorch_lightning.py | 0.923338 | 0.305166 | pytorch_lightning.py | pypi |
from typing import Dict, List, Union, Callable, Optional
from ray import tune
import os
from ray.tune.utils import flatten_dict
from lightgbm.callback import CallbackEnv
from lightgbm.basic import Booster
class TuneCallback:
"""Base class for Tune's LightGBM callbacks."""
pass
class TuneReportCallback(Tu... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/lightgbm.py | 0.966291 | 0.270827 | lightgbm.py | pypi |
import os
from typing import Any, Optional, Tuple, List
import subprocess
from ray import logger
from ray.autoscaler._private.command_runner import KubernetesCommandRunner
from ray.tune.syncer import NodeSyncer
from ray.tune.sync_client import SyncClient
from ray.util import get_node_ip_address
def try_import_kubern... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/kubernetes.py | 0.884214 | 0.183319 | kubernetes.py | pypi |
from typing import Callable, Dict, List, Union, Optional
from collections import OrderedDict
from ray import tune
import os
from ray.tune.utils import flatten_dict
from xgboost.core import Booster
try:
from xgboost.callback import TrainingCallback
except ImportError:
class TrainingCallback:
pass
c... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/xgboost.py | 0.97377 | 0.304688 | xgboost.py | pypi |
from contextlib import contextmanager
import os
import logging
import shutil
import tempfile
from typing import Callable, Dict, Generator, Optional, Type
import torch
from datetime import timedelta
import ray
from ray import tune
from ray.tune.result import RESULT_DUPLICATE
from ray.tune.logger import NoopLogger
from... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/torch.py | 0.845751 | 0.238694 | torch.py | pypi |
from typing import Callable, Dict, Type, Optional
from contextlib import contextmanager
import os
import logging
import shutil
import tempfile
from filelock import FileLock
import ray
from ray import tune
from ray.tune.function_runner import wrap_function
from ray.tune.logger import NoopLogger
from ray.tune.result i... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/horovod.py | 0.891628 | 0.250511 | horovod.py | pypi |
from typing import Dict, List, Union
from ray import tune
import mxnet
from mxnet.model import save_checkpoint, BatchEndParam
import numpy as np
import os
class TuneCallback:
"""Base class for Tune's MXNet callbacks."""
pass
class TuneReportCallback(TuneCallback):
"""MXNet to Ray Tune reporting call... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/mxnet.py | 0.955413 | 0.288929 | mxnet.py | pypi |
import json
import logging
import ray
import os
from ray.util.ml_utils.util import find_free_port
from ray.tune.function_runner import wrap_function
from ray.tune.resources import Resources
from ray.tune.result import RESULT_DUPLICATE
from ray.tune.trainable import DistributedTrainable
from ray.tune.utils import detec... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/integration/tensorflow.py | 0.749729 | 0.327507 | tensorflow.py | pypi |
import copy
import os
from typing import Any, Callable, Dict, Optional, Type, Union
import ray.cloudpickle as pickle
from ray.ml.config import RunConfig
from ray.ml.trainer import Trainer
from ray.tune import Experiment, TuneError, ExperimentAnalysis
from ray.tune.impl.utils import execute_dataset
from ray.tune.result... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/impl/tuner_internal.py | 0.900351 | 0.216094 | tuner_internal.py | pypi |
from django.db import models
class JobRecord(models.Model):
"""Information of an AutoML Job."""
job_id = models.CharField(max_length=50)
name = models.CharField(max_length=20)
user = models.CharField(max_length=20)
type = models.CharField(max_length=20)
start_time = models.CharField(max_lengt... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/models/models.py | 0.828523 | 0.231213 | models.py | pypi |
from django.shortcuts import HttpResponse
from ray.tune.automlboard.models.models import JobRecord, TrialRecord
from ray.tune.trial import Trial
import json
def query_job(request):
"""Rest API to query the job info, with the given job_id.
The url pattern should be like this:
curl http://<server>:<port... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/frontend/query.py | 0.48121 | 0.322259 | query.py | pypi |
import logging
import os
import time
from threading import Thread
from ray.tune.automlboard.common.exception import CollectorError
from ray.tune.automlboard.common.utils import (
parse_json,
parse_multiple_json,
timestamp2date,
)
from ray.tune.automlboard.models.models import JobRecord, TrialRecord, Resul... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/backend/collector.py | 0.744563 | 0.158435 | collector.py | pypi |
import logging
import json
import os
import time
def dump_json(json_info, json_file, overwrite=True):
"""Dump a whole json record into the given file.
Overwrite the file if the overwrite flag set.
Args:
json_info (dict): Information dict to be dumped.
json_file (str): File path to be dum... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automlboard/common/utils.py | 0.563138 | 0.185007 | utils.py | pypi |
import json
import logging
import os
import warnings
import traceback
from numbers import Number
from typing import Any, Dict, List, Optional, Tuple
from ray.ml.checkpoint import Checkpoint
from ray.tune.cloud import TrialCheckpoint
from ray.util.debug import log_once
from ray.tune.syncer import SyncConfig
from ray.tu... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/analysis/experiment_analysis.py | 0.892475 | 0.182098 | experiment_analysis.py | pypi |
# __import_lightning_begin__
import math
import torch
import pytorch_lightning as pl
from filelock import FileLock
from torch.utils.data import DataLoader, random_split
from torch.nn import functional as F
from torchvision.datasets import MNIST
from torchvision import transforms
import os
# __import_lightning_end__
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_pytorch_lightning.py | 0.880271 | 0.439687 | mnist_pytorch_lightning.py | pypi |
import mxnet as mx
from ray import tune, logger
from ray.tune.integration.mxnet import TuneCheckpointCallback, TuneReportCallback
from ray.tune.schedulers import ASHAScheduler
def train_mnist_mxnet(config, mnist, num_epochs=10):
batch_size = config["batch_size"]
train_iter = mx.io.NDArrayIter(
mnist[... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mxnet_example.py | 0.730386 | 0.406332 | mxnet_example.py | pypi |
import argparse
import tensorflow as tf
import numpy as np
import ray
from ray import tune
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.integration.keras import TuneReportCheckpointCallback
from ray.tune.integration.tensorflow import DistributedTrainableCreator, get_num_workers
def mnist_data... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tf_distributed_keras_example.py | 0.848831 | 0.460774 | tf_distributed_keras_example.py | pypi |
import argparse
import os
from filelock import FileLock
from tensorflow.keras.layers import Dense, Flatten, Conv2D
from tensorflow.keras import Model
from tensorflow.keras.datasets.mnist import load_data
from ray import tune
MAX_TRAIN_BATCH = 10
class MyModel(Model):
def __init__(self, hiddens=128):
s... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tf_mnist_example.py | 0.903664 | 0.407098 | tf_mnist_example.py | pypi |
import argparse
import json
import time
import os
import numpy as np
import ray
from ray import tune
from ray.tune import Trainable
from ray.tune.schedulers.hb_bohb import HyperBandForBOHB
from ray.tune.suggest.bohb import TuneBOHB
class MyTrainableClass(Trainable):
"""Example agent whose learning curve is a r... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/bohb_example.py | 0.850531 | 0.307267 | bohb_example.py | pypi |
import torch
import torch.nn as nn
import os
import numpy as np
import torchvision
from torch.utils.data import DataLoader
import torchvision.transforms as transforms
import ray
from ray import tune
from ray.tune.schedulers import create_scheduler
from ray.tune.integration.horovod import (
DistributedTrainableCre... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/horovod_cifar_pbt_example.py | 0.757615 | 0.418875 | horovod_cifar_pbt_example.py | pypi |
import time
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.optuna import OptunaSearch
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objectiv... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_example.py | 0.753194 | 0.273485 | optuna_example.py | pypi |
import math
import torch
from filelock import FileLock
from torch.nn import functional as F
from torchmetrics import Accuracy
import pytorch_lightning as pl
from pl_bolts.datamodules.mnist_datamodule import MNISTDataModule
import os
from ray.tune.integration.pytorch_lightning import TuneReportCallback
from ray import... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_ptl_mini.py | 0.902742 | 0.47859 | mnist_ptl_mini.py | pypi |
from __future__ import print_function
import argparse
import os
import torch
import torch.optim as optim
import ray
from ray import tune
from ray.tune.schedulers import ASHAScheduler
from ray.tune.examples.mnist_pytorch import train, test, get_data_loaders, ConvNet
# Change these values if you want the training to r... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mnist_pytorch_trainable.py | 0.852935 | 0.278272 | mnist_pytorch_trainable.py | pypi |
import time
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.skopt import SkOptSearch
def evaluation_fn(step, width, height):
time.sleep(0.1)
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_o... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/skopt_example.py | 0.811713 | 0.287565 | skopt_example.py | pypi |
import time
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.flaml import BlendSearch
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objective(... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/blendsearch_example.py | 0.74008 | 0.317942 | blendsearch_example.py | pypi |
# __tutorial_imports_begin__
import argparse
import os
import numpy as np
import torch
import torch.optim as optim
from ray.tune.examples.mnist_pytorch import train, test, ConvNet, get_data_loaders
from ray import tune
from ray.tune.schedulers import PopulationBasedTraining
from ray.tune.trial import ExportFormat
# ... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_convnet_function_example.py | 0.870184 | 0.38659 | pbt_convnet_function_example.py | pypi |
from __future__ import print_function
from tensorflow.keras.models import Sequential, Model, load_model
from tensorflow.keras.layers import Embedding
from tensorflow.keras.layers import Input, Activation, Dense, Permute, Dropout
from tensorflow.keras.layers import add, dot, concatenate
from tensorflow.keras.layers imp... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_memnn_example.py | 0.715325 | 0.366391 | pbt_memnn_example.py | pypi |
from typing import Dict, List
import sklearn.datasets
import sklearn.metrics
import os
import numpy as np
from ray.tune.schedulers import ASHAScheduler
from sklearn.model_selection import train_test_split
import xgboost as xgb
from ray import tune
from ray.tune.integration.xgboost import (
TuneReportCheckpointCall... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/xgboost_example.py | 0.903018 | 0.550003 | xgboost_example.py | pypi |
import random
from ray import tune
from ray.tune.schedulers import PopulationBasedTraining
if __name__ == "__main__":
# Postprocess the perturbed config to ensure it's still valid
def explore(config):
# ensure we collect enough timesteps to do sgd
if config["train_batch_size"] < config["sgd_m... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_ppo_example.py | 0.675015 | 0.261985 | pbt_ppo_example.py | pypi |
import argparse
import time
import ray
from ray import tune
from ray.tune.schedulers import AsyncHyperBandScheduler
def evaluation_fn(step, width, height):
time.sleep(0.1)
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objective(config):
# Hyperparameters
width, height = config[... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/async_hyperband_example.py | 0.731826 | 0.268216 | async_hyperband_example.py | pypi |
import time
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.flaml import CFO
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objective(config):... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/cfo_example.py | 0.752013 | 0.28208 | cfo_example.py | pypi |
import sys
import time
from ray import tune
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.sigopt import SigOptSearch
def evaluate(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.01
def easy_objective(config):
# Hyperparameters
width, height =... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/sigopt_example.py | 0.583085 | 0.261213 | sigopt_example.py | pypi |
"""Examples using MLfowLoggerCallback and mlflow_mixin.
"""
import os
import tempfile
import time
import mlflow
from ray import tune
from ray.tune.integration.mlflow import MLflowLoggerCallback, mlflow_mixin
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def e... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mlflow_example.py | 0.729905 | 0.471467 | mlflow_example.py | pypi |
import os
import time
import json
import argparse
from ray import tune
def evaluation_fn(step, width, height):
time.sleep(0.1)
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def train_func(config, checkpoint_dir=None):
start = 0
width, height = config["width"], config["height"]
if ch... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/custom_func_checkpointing.py | 0.669205 | 0.174621 | custom_func_checkpointing.py | pypi |
import argparse
import numpy as np
import time
import logging
import os
import ray
from ray import tune
from ray.tune import Trainable
from ray.tune.sync_client import get_sync_client
from ray import cloudpickle
logger = logging.getLogger(__name__)
class MockDurableTrainable(Trainable):
"""Mocks the storage cli... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/durable_trainable_example.py | 0.753648 | 0.218732 | durable_trainable_example.py | pypi |
# __import_begin__
from functools import partial
import numpy as np
import os
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from filelock import FileLock
from torch.utils.data import random_split
import torchvision
import torchvision.transforms as transforms
import ray
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/cifar10_pytorch.py | 0.888819 | 0.327897 | cifar10_pytorch.py | pypi |
import argparse
import json
import os
import numpy as np
import ray
from ray import tune
from ray.tune.schedulers import HyperBandScheduler
def train(config, checkpoint_dir=None):
step = 0
if checkpoint_dir:
with open(os.path.join(checkpoint_dir, "checkpoint")) as f:
step = json.loads(f... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/hyperband_function_example.py | 0.656988 | 0.224523 | hyperband_function_example.py | pypi |
import numpy as np
import time
from ray import tune
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.ax import AxSearch
def hartmann6(x):
alpha = np.array([1.0, 1.2, 3.0, 3.2])
A = np.array(
[
[10, 3, 17, 3.5, 1.7, 8],
[0.05, 10, 17, 0.1, 8, 14],
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/ax_example.py | 0.646795 | 0.393356 | ax_example.py | pypi |
import argparse
import os
from filelock import FileLock
from tensorflow.keras.datasets import mnist
import ray
from ray import tune
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.integration.keras import TuneReportCallback
def train_mnist(config):
# https://github.com/tensorflow/tensorflo... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tune_mnist_keras.py | 0.812459 | 0.379206 | tune_mnist_keras.py | pypi |
# flake8: noqa
# fmt: off
# __tutorial_imports_begin__
import argparse
import os
import numpy as np
import torch
import torch.optim as optim
from torchvision import datasets
from ray.tune.examples.mnist_pytorch import train, test, ConvNet,\
get_data_loaders
import ray
from ray import tune
from ray.tune.scheduler... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_convnet_example.py | 0.830285 | 0.464841 | pbt_convnet_example.py | pypi |
import time
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.nevergrad import NevergradSearch
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objective(con... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/nevergrad_example.py | 0.791821 | 0.274572 | nevergrad_example.py | pypi |
import argparse
import logging
import os
import torch
import torch.optim as optim
from torch.nn.parallel import DistributedDataParallel
import ray
from ray import tune
from ray.tune.examples.mnist_pytorch import train, test, get_data_loaders, ConvNet
from ray.tune.integration.torch import (
DistributedTrainableCre... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/ddp_mnist_torch.py | 0.734024 | 0.310459 | ddp_mnist_torch.py | pypi |
HPO, and MLflow autologging all together."""
import os
import tempfile
import pytorch_lightning as pl
from pl_bolts.datamodules import MNISTDataModule
import mlflow
from ray import tune
from ray.tune.integration.mlflow import mlflow_mixin
from ray.tune.integration.pytorch_lightning import TuneReportCallback
from ray... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/mlflow_ptl.py | 0.687525 | 0.535645 | mlflow_ptl.py | pypi |
import time
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.hyperopt import HyperOptSearch
from hyperopt import hp
def f_unpack_dict(dct):
"""
Unpacks all sub-dictionaries in given dictionary recursi... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/hyperopt_conditional_search_space_example.py | 0.905034 | 0.312003 | hyperopt_conditional_search_space_example.py | pypi |
import time
from typing import Dict, Optional, Any
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.optuna import OptunaSearch
def evaluation_fn(step, width, height, mult=1):
return (0.1 + width * step / ... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_define_by_run_example.py | 0.920598 | 0.350616 | optuna_define_by_run_example.py | pypi |
import argparse
import tempfile
from unittest.mock import MagicMock
import numpy as np
import wandb
from ray import tune
from ray.tune import Trainable
from ray.tune.integration.wandb import (
WandbLoggerCallback,
WandbTrainableMixin,
wandb_mixin,
)
def train_function(config, checkpoint_dir=None):
f... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/wandb_example.py | 0.771628 | 0.409693 | wandb_example.py | pypi |
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
import time
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.schedulers import AsyncHyperBandScheduler
from ray.tune.suggest.dragonfly import DragonflySearc... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/dragonfly_example.py | 0.729134 | 0.19095 | dragonfly_example.py | pypi |
from __future__ import print_function
import argparse
import random
import mxnet as mx
import numpy as np
from mxnet import gluon, init
from mxnet import autograd as ag
from mxnet.gluon import nn
from mxnet.gluon.data.vision import transforms
from gluoncv.model_zoo import get_model
from gluoncv.data import transform... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/tune_cifar10_gluon.py | 0.873768 | 0.264142 | tune_cifar10_gluon.py | pypi |
import time
import ray
from ray import tune
from ray.tune.suggest import ConcurrencyLimiter
from ray.tune.suggest.optuna import OptunaSearch
def evaluation_fn(step, width, height):
return (0.1 + width * step / 100) ** (-1) + height * 0.1
def easy_objective(config):
# Hyperparameters
width, height = con... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/optuna_multiobjective_example.py | 0.732783 | 0.272464 | optuna_multiobjective_example.py | pypi |
from __future__ import print_function
import argparse
import numpy as np
import tensorflow as tf
from tensorflow.keras.datasets import cifar10
from tensorflow.keras.layers import Input, Dense, Dropout, Flatten
from tensorflow.keras.layers import Convolution2D, MaxPooling2D
from tensorflow.keras.models import Model, l... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_tune_cifar10_with_keras.py | 0.9151 | 0.443721 | pbt_tune_cifar10_with_keras.py | pypi |
import os
import ray
from ray import tune
from ray.tune import CLIReporter
from ray.tune.examples.pbt_transformers.utils import (
download_data,
build_compute_metrics_fn,
)
from ray.tune.schedulers import PopulationBasedTraining
from transformers import (
glue_tasks_num_labels,
AutoConfig,
AutoMode... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/examples/pbt_transformers/pbt_transformers.py | 0.578805 | 0.258431 | pbt_transformers.py | pypi |
from typing import Dict, Optional
from copy import deepcopy
import logging
import numpy as np
import pandas as pd
from ray.tune import TuneError
from ray.tune.schedulers import PopulationBasedTraining
def import_pb2_dependencies():
try:
import GPy
except ImportError:
GPy = None
try:
... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/pb2.py | 0.849909 | 0.58255 | pb2.py | pypi |
import collections
import logging
from typing import Dict, List, Optional
import numpy as np
from ray.tune import trial_runner
from ray.tune.result import DEFAULT_METRIC
from ray.tune.trial import Trial
from ray.tune.schedulers.trial_scheduler import FIFOScheduler, TrialScheduler
logger = logging.getLogger(__name__)... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/median_stopping_rule.py | 0.943073 | 0.345202 | median_stopping_rule.py | pypi |
import numpy as np
from scipy.optimize import minimize
import GPy
from GPy.kern import Kern
from GPy.core import Param
from sklearn.metrics import pairwise_distances
from sklearn.metrics.pairwise import euclidean_distances
class TV_SquaredExp(Kern):
"""Time varying squared exponential kernel.
For more info s... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/pb2_utils.py | 0.868715 | 0.614076 | pb2_utils.py | pypi |
from typing import Dict, Optional
from ray.tune import trial_runner
from ray.tune.result import DEFAULT_METRIC
from ray.tune.trial import Trial
class TrialScheduler:
"""Interface for implementing a Trial Scheduler class."""
CONTINUE = "CONTINUE" #: Status for continuing trial execution
PAUSE = "PAUSE" ... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/trial_scheduler.py | 0.883645 | 0.24164 | trial_scheduler.py | pypi |
from ray._private.utils import get_function_args
from ray.tune.schedulers.trial_scheduler import TrialScheduler, FIFOScheduler
from ray.tune.schedulers.hyperband import HyperBandScheduler
from ray.tune.schedulers.hb_bohb import HyperBandForBOHB
from ray.tune.schedulers.async_hyperband import AsyncHyperBandScheduler, AS... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/schedulers/__init__.py | 0.905387 | 0.440469 | __init__.py | pypi |
import time
import copy
import logging
from ray.tune.trial import Trial
from ray.tune.suggest import SearchAlgorithm
from ray.tune.experiment import convert_to_experiment_list
from ray.tune.suggest.variant_generator import generate_variants
from ray.tune.config_parser import make_parser, create_trial_from_spec
logger... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automl/search_policy.py | 0.713531 | 0.192748 | search_policy.py | pypi |
import random
import logging
import numpy as np
from ray.tune import grid_search
logger = logging.getLogger(__name__)
class ParameterSpace:
"""Base class of a single parameter's search space."""
def __init__(self, name):
"""Initialize ParameterSpace.
Arguments:
name (str): Name... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/automl/search_space.py | 0.787155 | 0.457076 | search_space.py | pypi |
from typing import Dict, List, Union
import copy
import glob
import logging
import os
import inspect
import threading
import time
import uuid
from collections import defaultdict
from datetime import datetime
from threading import Thread
from typing import Optional
import numpy as np
import ray
import psutil
from ray.... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/utils/util.py | 0.813535 | 0.154695 | util.py | pypi |
from typing import Optional, Dict, Any
import logging
import os
import time
import ray
from ray import ray_constants
from ray._private.resource_spec import NODE_ID_PREFIX
from ray.tune.resources import Resources
logger = logging.getLogger(__name__)
TUNE_STATE_REFRESH_PERIOD = 10 # Refresh resources every 10 s
def... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/tune/utils/resource_updater.py | 0.888457 | 0.170128 | resource_updater.py | pypi |
from typing import Any, List, Tuple, Dict, Optional
class CommandRunnerInterface:
"""Interface to run commands on a remote cluster node.
**Important**: This is an INTERNAL API that is only exposed for the purpose
of implementing custom node providers. It is not allowed to call into
CommandRunner meth... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/command_runner.py | 0.916717 | 0.310707 | command_runner.py | pypi |
import logging
from types import ModuleType
from typing import Any, Dict, List, Optional
from ray.autoscaler.command_runner import CommandRunnerInterface
from ray.autoscaler._private.command_runner import SSHCommandRunner, DockerCommandRunner
logger = logging.getLogger(__name__)
class NodeProvider:
"""Interface... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/node_provider.py | 0.941654 | 0.339828 | node_provider.py | pypi |
from enum import Enum, auto
from typing import Any, Callable, Dict, List, Optional, Union
from ray.autoscaler._private.cli_logger import cli_logger
class CreateClusterEvent(Enum):
"""Events to track in ray.autoscaler.sdk.create_or_update_cluster.
Attributes:
up_started : Invoked at the beginning of ... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/_private/event_system.py | 0.909959 | 0.219578 | event_system.py | pypi |
from ray.autoscaler._private import constants
from typing import List, Set, Tuple
class NodeTracker:
"""Map nodes to their corresponding logs.
We need to be a little careful here. At an given point in time, node_id <->
ip can be interchangeably used, but the node_id -> ip relation is not
bijective _a... | /ray_for_mars-1.12.1-cp38-cp38-manylinux2014_x86_64.whl/ray_for_mars-1.12.1.data/purelib/ray/autoscaler/_private/node_tracker.py | 0.911031 | 0.482673 | node_tracker.py | pypi |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.