after_merge stringlengths 64 17k | before_merge stringlengths 60 17k | source code and errors stringlengths 236 32.3k | full_traceback stringlengths 170 17.7k | traceback_type stringclasses 60
values |
|---|---|---|---|---|
def _blocking_send(self, req: ray_client_pb2.DataRequest
) -> ray_client_pb2.DataResponse:
req_id = self._next_id()
req.req_id = req_id
self.request_queue.put(req)
data = None
with self.cv:
self.cv.wait_for(
lambda: req_id in... | def _blocking_send(self, req: ray_client_pb2.DataRequest
) -> ray_client_pb2.DataResponse:
req_id = self._next_id()
req.req_id = req_id
self.request_queue.put(req)
data = None
with self.cv:
self.cv.wait_for(lambda: req_id in self.ready_data)... | [{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive... | Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Received http2 header with status: 502"
debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr... | debug_error |
def _log_main(self) -> None:
stub = ray_client_pb2_grpc.RayletLogStreamerStub(self.channel)
log_stream = stub.Logstream(
iter(self.request_queue.get, None), metadata=self._metadata)
try:
for record in log_stream:
if record.level < 0:
... | def _log_main(self) -> None:
stub = ray_client_pb2_grpc.RayletLogStreamerStub(self.channel)
log_stream = stub.Logstream(
iter(self.request_queue.get, None), metadata=self._metadata)
try:
for record in log_stream:
if record.level < 0:
... | [{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive... | Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Received http2 header with status: 502"
debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr... | debug_error |
def __init__(self,
conn_str: str = "",
secure: bool = False,
metadata: List[Tuple[str, str]] = None,
connection_retries: int = 3):
"""Initializes the worker side grpc client.
Args:
conn_str: The host:port connection str... | def __init__(self,
conn_str: str = "",
secure: bool = False,
metadata: List[Tuple[str, str]] = None,
connection_retries: int = 3):
"""Initializes the worker side grpc client.
Args:
conn_str: The host:port connection str... | [{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive... | Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Received http2 header with status: 502"
debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr... | debug_error |
def translate(configuration: Dict[str, Any],
dictionary: Dict[str, str]) -> Dict[str, Any]:
return {
dictionary[field]: configuration[field]
for field in dictionary if field in configuration
} | def translate(configuration: Dict[str, Any],
dictionary: Dict[str, str]) -> Dict[str, Any]:
return {dictionary[field]: configuration[field] for field in dictionary} | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ray/anaconda3/bin/ray-operator", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ray/anaconda3/lib/python3.7/site-packages/ray/operator/operator.py", line 123, in main\\ncluster_config = operator_utils.cr_to_config(... | Traceback (most recent call last):
File "/home/ray/anaconda3/bin/ray-operator", line 8, in <module>
sys.exit(main())
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/operator/operator.py", line 123, in main
cluster_config = operator_utils.cr_to_config(cluster_cr)
File "/home/ray/anaconda3/lib/python3.7/site-pa... | KeyError |
def __init__(self,
conn_str: str = "",
secure: bool = False,
metadata: List[Tuple[str, str]] = None,
connection_retries: int = 3):
"""Initializes the worker side grpc client.
Args:
conn_str: The host:port connection str... | def __init__(self,
conn_str: str = "",
secure: bool = False,
metadata: List[Tuple[str, str]] = None,
connection_retries: int = 3):
"""Initializes the worker side grpc client.
Args:
conn_str: The host:port connection str... | [{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610606646.739755640","description":"Receive... | Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "Received http2 header with status: 502"
debug_error_string = "{"created":"@1610606646.739755640","description":"Received http2 :status header with non-200 OK status","file":"sr... | debug_error |
def __init__(self,
name: str,
description: str = "",
tag_keys: Optional[Tuple[str]] = None):
if len(name) == 0:
raise ValueError("Empty name is not allowed. "
"Please provide a metric name.")
self._name = nam... | def __init__(self,
name: str,
description: str = "",
tag_keys: Optional[Tuple[str]] = None):
if len(name) == 0:
raise ValueError("Empty name is not allowed. "
"Please provide a metric name.")
self._name = nam... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ... | Traceback (most recent call last):
File "test.py", line 8, in <module>
count.record(1.0, {"b": 2})
File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record
self._metric.record(value, tags=final_tags)
File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record
c_tags[tag_k.encode("as... | AttributeError |
def set_default_tags(self, default_tags: Dict[str, str]):
"""Set default tags of metrics.
Example:
>>> # Note that set_default_tags returns the instance itself.
>>> counter = Counter("name")
>>> counter2 = counter.set_default_tags({"a": "b"})
>>> asse... | def set_default_tags(self, default_tags: Dict[str, str]):
"""Set default tags of metrics.
Example:
>>> # Note that set_default_tags returns the instance itself.
>>> counter = Counter("name")
>>> counter2 = counter.set_default_tags({"a": "b"})
>>> asse... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ... | Traceback (most recent call last):
File "test.py", line 8, in <module>
count.record(1.0, {"b": 2})
File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record
self._metric.record(value, tags=final_tags)
File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record
c_tags[tag_k.encode("as... | AttributeError |
def record(self, value: float, tags: dict = None) -> None:
"""Record the metric point of the metric.
Args:
value(float): The value to be recorded as a metric point.
"""
assert self._metric is not None
if tags is not None:
for val in tags.values():
... | def record(self, value: float, tags: dict = None) -> None:
"""Record the metric point of the metric.
Args:
value(float): The value to be recorded as a metric point.
"""
assert self._metric is not None
default_tag_copy = self._default_tags.copy()
default_t... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ... | Traceback (most recent call last):
File "test.py", line 8, in <module>
count.record(1.0, {"b": 2})
File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record
self._metric.record(value, tags=final_tags)
File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record
c_tags[tag_k.encode("as... | AttributeError |
def postprocess_advantages(policy,
sample_batch,
other_agent_batches=None,
episode=None):
# Stub serving backward compatibility.
deprecation_warning(
old="rllib.agents.a3c.a3c_tf_policy.postprocess_advantages",
new... | def postprocess_advantages(policy,
sample_batch,
other_agent_batches=None,
episode=None):
completed = sample_batch[SampleBatch.DONES][-1]
if completed:
last_r = 0.0
else:
next_state = []
for i in range(p... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def setup_mixins(policy, obs_space, action_space, config):
ValueNetworkMixin.__init__(policy, obs_space, action_space, config)
LearningRateSchedule.__init__(policy, config["lr"], config["lr_schedule"]) | def setup_mixins(policy, obs_space, action_space, config):
ValueNetworkMixin.__init__(policy)
LearningRateSchedule.__init__(policy, config["lr"], config["lr_schedule"]) | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def __init__(self,
action_dist,
actions,
advantages,
v_target,
vf,
vf_loss_coeff=0.5,
entropy_coeff=0.01):
log_prob = action_dist.logp(actions)
# The "policy gradients" loss
... | def __init__(self):
@make_tf_callable(self.get_session())
def value(ob, prev_action, prev_reward, *state):
model_out, _ = self.model({
SampleBatch.CUR_OBS: tf.convert_to_tensor([ob]),
SampleBatch.PREV_ACTIONS: tf.convert_to_tensor([prev_action]),
... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def add_advantages(policy,
sample_batch,
other_agent_batches=None,
episode=None):
# Stub serving backward compatibility.
deprecation_warning(
old="rllib.agents.a3c.a3c_torch_policy.add_advantages",
new="rllib.evaluation.postprocessing.com... | def add_advantages(policy,
sample_batch,
other_agent_batches=None,
episode=None):
completed = sample_batch[SampleBatch.DONES][-1]
if completed:
last_r = 0.0
else:
last_r = policy._value(sample_batch[SampleBatch.NEXT_OBS][-1])
ret... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def clip_gradients(policy: Policy, optimizer: "tf.keras.optimizers.Optimizer",
loss: TensorType) -> ModelGradients:
return minimize_and_clip(
optimizer,
loss,
var_list=policy.q_func_vars,
clip_val=policy.config["grad_clip"]) | def clip_gradients(policy: Policy, optimizer: "tf.keras.optimizers.Optimizer",
loss: TensorType) -> ModelGradients:
if policy.config["grad_clip"] is not None:
grads_and_vars = minimize_and_clip(
optimizer,
loss,
var_list=policy.q_func_vars,
... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def get_policy_class(config):
if config["framework"] == "torch":
if config["vtrace"]:
from ray.rllib.agents.impala.vtrace_torch_policy import \\
VTraceTorchPolicy
return VTraceTorchPolicy
else:
from ray.rllib.agents.a3c.a3c_torch_policy import \\
... | def get_policy_class(config):
if config["framework"] == "torch":
if config["vtrace"]:
from ray.rllib.agents.impala.vtrace_torch_policy import \\
VTraceTorchPolicy
return VTraceTorchPolicy
else:
from ray.rllib.agents.a3c.a3c_torch_policy import \\
... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def postprocess_trajectory(
policy: Policy,
sample_batch: SampleBatch,
other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None,
episode: Optional[MultiAgentEpisode] = None) -> SampleBatch:
"""Postprocesses a trajectory and returns the processed trajectory.
The trajector... | def postprocess_trajectory(
policy: Policy,
sample_batch: SampleBatch,
other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None,
episode: Optional[MultiAgentEpisode] = None) -> SampleBatch:
"""Postprocesses a trajectory and returns the processed trajectory.
The trajector... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def postprocess_ppo_gae(
policy: Policy,
sample_batch: SampleBatch,
other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None,
episode: Optional[MultiAgentEpisode] = None) -> SampleBatch:
# Stub serving backward compatibility.
deprecation_warning(
old="rllib.agent... | def postprocess_ppo_gae(
policy: Policy,
sample_batch: SampleBatch,
other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None,
episode: Optional[MultiAgentEpisode] = None) -> SampleBatch:
"""Postprocesses a trajectory and returns the processed trajectory.
The trajectory c... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def gradients(self, optimizer, loss):
self.gvs = {
k: minimize_and_clip(optimizer, self.losses[k], self.vars[k],
self.config["grad_norm_clipping"])
for k, optimizer in self.optimizers.items()
}
return self.gvs["critic"] + self.gvs["act... | def gradients(self, optimizer, loss):
if self.config["grad_norm_clipping"] is not None:
self.gvs = {
k: minimize_and_clip(optimizer, self.losses[k], self.vars[k],
self.config["grad_norm_clipping"])
for k, optimizer in self.opti... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def minimize_and_clip(optimizer, objective, var_list, clip_val=10.0):
"""Minimized `objective` using `optimizer` w.r.t. variables in
`var_list` while ensure the norm of the gradients for each
variable is clipped to `clip_val`
"""
# Accidentally passing values < 0.0 will break all gradients.
asse... | def minimize_and_clip(optimizer, objective, var_list, clip_val=10.0):
"""Minimized `objective` using `optimizer` w.r.t. variables in
`var_list` while ensure the norm of the gradients for each
variable is clipped to `clip_val`
"""
# Accidentally passing values < 0.0 will break all gradients.
asse... | [{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro... | 2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.
Traceback (most recent call last):
File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial
result = self.trial_executor.fetch_result(tria... | tensorflow.python.framework.errors_impl.InvalidArgumentError |
def persistent_id(self, obj):
if isinstance(obj, ray.ObjectRef):
obj_id = obj.binary()
if obj_id not in self.server.object_refs[self.client_id]:
# We're passing back a reference, probably inside a reference.
# Let's hold onto it.
self.s... | def persistent_id(self, obj):
if isinstance(obj, ray.ObjectRef):
obj_id = obj.binary()
if obj_id not in self.server.object_refs[self.client_id]:
# We're passing back a reference, probably inside a reference.
# Let's hold onto it.
self.s... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\n@ray.remote\\nclass B:\\npass\\n\\n@ray.remote\\nclass A:\\ndef __init__(self):\\nself.b = B.remote()\\n\\ndef get_actor_ref(self):\\nreturn [self.b]\\n\\na = A.remote()\\nray.get(a.get_actor_ref.remote())'}, {'piece_... | Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.UNKNOWN
details = "Exception iterating responses: 'ServerPickler' object has no attribute 'actor_refs'"
debug_error_string = "{"created":"@1610645797.119742000","description":"Error received from pee... | debug_error |
def ClusterInfo(self, request,
context=None) -> ray_client_pb2.ClusterInfoResponse:
resp = ray_client_pb2.ClusterInfoResponse()
resp.type = request.type
if request.type == ray_client_pb2.ClusterInfoType.CLUSTER_RESOURCES:
with disable_client_hook():
... | def ClusterInfo(self, request,
context=None) -> ray_client_pb2.ClusterInfoResponse:
resp = ray_client_pb2.ClusterInfoResponse()
resp.type = request.type
if request.type == ray_client_pb2.ClusterInfoType.CLUSTER_RESOURCES:
with disable_client_hook():
... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\nprint(ray.get_runtime_context().node_id.hex())'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "test.py", line 4, in <module>\\nprint(ray.get_runtime_context().node_id.he... | Traceback (most recent call last):
File "test.py", line 4, in <module>
print(ray.get_runtime_context().node_id.hex())
File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 58, in node_id
node_id = self.worker.current_node_id
File "/Users/eoakes/code/ray/python/ray/worker.py", line 148, in current_node_id
re... | AttributeError |
def get_cluster_info(self, type: ray_client_pb2.ClusterInfoType.TypeEnum):
req = ray_client_pb2.ClusterInfoRequest()
req.type = type
resp = self.server.ClusterInfo(req, metadata=self.metadata)
if resp.WhichOneof("response_type") == "resource_table":
# translate from a pro... | def get_cluster_info(self, type: ray_client_pb2.ClusterInfoType.TypeEnum):
req = ray_client_pb2.ClusterInfoRequest()
req.type = type
resp = self.server.ClusterInfo(req, metadata=self.metadata)
if resp.WhichOneof("response_type") == "resource_table":
# translate from a pro... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\nprint(ray.get_runtime_context().node_id.hex())'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "test.py", line 4, in <module>\\nprint(ray.get_runtime_context().node_id.he... | Traceback (most recent call last):
File "test.py", line 4, in <module>
print(ray.get_runtime_context().node_id.hex())
File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 58, in node_id
node_id = self.worker.current_node_id
File "/Users/eoakes/code/ray/python/ray/worker.py", line 148, in current_node_id
re... | AttributeError |
def get(self):
"""Get a dictionary of the current context.
Fields that are not available (e.g., actor ID inside a task) won't be
included in the field.
Returns:
dict: Dictionary of the current context.
"""
context = {
"job_id": self.job_id,
... | def get(self):
"""Get a dictionary of the current_context.
For fields that are not available (for example actor id inside a task)
won't be included in the field.
Returns:
dict: Dictionary of the current context.
"""
context = {
"job_id": self... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nray.init()\\n\\nprint(ray.get_runtime_context().get())'}, {'piece_type': 'error message', 'piece_content': '2021-01-13 14:15:29,011 INFO services.py:1169 -- View the Ray dashboard at http://127.0.0.1:8266\\nTraceback (most recent call last):\\nFile "test.py"... | 2021-01-13 14:15:29,011 INFO services.py:1169 -- View the Ray dashboard at http://127.0.0.1:8266
Traceback (most recent call last):
File "test.py", line 4, in <module>
print(ray.get_runtime_context().get())
File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 26, in get
"task_id": self.task_id,
File "/User... | AssertionError |
def __init__(self,
node_ip_address,
redis_address,
dashboard_agent_port,
redis_password=None,
temp_dir=None,
log_dir=None,
metrics_export_port=None,
node_manager_port=None,
... | def __init__(self,
node_ip_address,
redis_address,
dashboard_agent_port,
redis_password=None,
temp_dir=None,
log_dir=None,
metrics_export_port=None,
node_manager_port=None,
... | [{'piece_type': 'error message', 'piece_content': '(pid=None) Traceback (most recent call last):\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\new_dashboard/agent.py", line 317, in <module>\\n(pid=None) raise e\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\e... | (pid=None) Traceback (most recent call last):
(pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 317, in <module>
(pid=None) raise e
(pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 293, in <modul... | KeyError |
async def run(self):
async def _check_parent():
"""Check if raylet is dead and fate-share if it is."""
try:
curr_proc = psutil.Process()
while True:
parent = curr_proc.parent()
if (parent is None or parent.pid ==... | async def run(self):
async def _check_parent():
"""Check if raylet is dead and fate-share if it is."""
try:
curr_proc = psutil.Process()
while True:
parent = curr_proc.parent()
if (parent is None or parent.pid ==... | [{'piece_type': 'error message', 'piece_content': '(pid=None) Traceback (most recent call last):\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\new_dashboard/agent.py", line 317, in <module>\\n(pid=None) raise e\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\e... | (pid=None) Traceback (most recent call last):
(pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 317, in <module>
(pid=None) raise e
(pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 293, in <modul... | KeyError |
def __str__(self):
return ("The worker died unexpectedly while executing this task. "
"Check python-core-worker-*.log files for more information.") | def __str__(self):
return "The worker died unexpectedly while executing this task." | [{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n... | 2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.
Traceback (most recent call last):
File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/usr... | ray.exceptions.RayTaskError |
def __str__(self):
return ("The actor died unexpectedly before finishing this task. "
"Check python-core-worker-*.log files for more information.") | def __str__(self):
return "The actor died unexpectedly before finishing this task." | [{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n... | 2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.
Traceback (most recent call last):
File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/usr... | ray.exceptions.RayTaskError |
def _try_to_compute_deterministic_class_id(cls, depth=5):
"""Attempt to produce a deterministic class ID for a given class.
The goal here is for the class ID to be the same when this is run on
different worker processes. Pickling, loading, and pickling again seems to
produce more consistent results tha... | def _try_to_compute_deterministic_class_id(cls, depth=5):
"""Attempt to produce a deterministic class ID for a given class.
The goal here is for the class ID to be the same when this is run on
different worker processes. Pickling, loading, and pickling again seems to
produce more consistent results tha... | [{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n... | 2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.
Traceback (most recent call last):
File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/usr... | ray.exceptions.RayTaskError |
def _random_string():
id_hash = hashlib.shake_128()
id_hash.update(uuid.uuid4().bytes)
id_bytes = id_hash.digest(ray_constants.ID_SIZE)
assert len(id_bytes) == ray_constants.ID_SIZE
return id_bytes | def _random_string():
id_hash = hashlib.sha1()
id_hash.update(uuid.uuid4().bytes)
id_bytes = id_hash.digest()
assert len(id_bytes) == ray_constants.ID_SIZE
return id_bytes | [{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n... | 2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.
Traceback (most recent call last):
File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/usr... | ray.exceptions.RayTaskError |
def run_function_on_all_workers(self, function,
run_on_other_drivers=False):
"""Run arbitrary code on all of the workers.
This function will first be run on the driver, and then it will be
exported to all of the workers to be run. It will also be run on a... | def run_function_on_all_workers(self, function,
run_on_other_drivers=False):
"""Run arbitrary code on all of the workers.
This function will first be run on the driver, and then it will be
exported to all of the workers to be run. It will also be run on a... | [{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n... | 2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.
Traceback (most recent call last):
File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/usr... | ray.exceptions.RayTaskError |
def _home(self):
if self._home_cached is not None:
return self._home_cached
for _ in range(MAX_HOME_RETRIES - 1):
try:
self._home_cached = self._try_to_get_home()
return self._home_cached
except Exception:
# TODO (Dm... | def _home(self):
# TODO (Dmitri): Think about how to use the node's HOME variable
# without making an extra kubectl exec call.
if self._home_cached is None:
cmd = self.kubectl + [
"exec", "-it", self.node_id, "--", "printenv", "HOME"
]
join... | [{'piece_type': 'error message', 'piece_content': 'Updating cluster configuration. [hash=0194a452ebd82e0ab6eade7b4dd3a4f4f775d5de]\\nNew status: syncing-files\\n[2/7] Processing file mounts\\n2020-12-15 09:03:24,116\\tINFO command_runner.py:169 -- NodeUpdater: ray-head-m5nvd: Running kubectl -n ray exec -it ray-head-m5... | Updating cluster configuration. [hash=0194a452ebd82e0ab6eade7b4dd3a4f4f775d5de]
New status: syncing-files
[2/7] Processing file mounts
2020-12-15 09:03:24,116 INFO command_runner.py:169 -- NodeUpdater: ray-head-m5nvd: Running kubectl -n ray exec -it ray-head-m5nvd -- bash --login -c -i 'true && source ~/.bashrc... | subprocess.CalledProcessError |
def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "debugging.py", line 2, in <module>\\nray.init()\\nFile "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init\\nray_params=ray_params)\\nFile "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init... | Traceback (most recent call last):
File "debugging.py", line 2, in <module>
ray.init()
File "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init
ray_params=ray_params)
File "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init__
self.start_head_processes()
File "/Users/haochen/code/ant_... | redis.exceptions.ResponseError |
def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "debugging.py", line 2, in <module>\\nray.init()\\nFile "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init\\nray_params=ray_params)\\nFile "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init... | Traceback (most recent call last):
File "debugging.py", line 2, in <module>
ray.init()
File "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init
ray_params=ray_params)
File "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init__
self.start_head_processes()
File "/Users/haochen/code/ant_... | redis.exceptions.ResponseError |
def shutdown(self) -> None:
"""Completely shut down the connected Serve instance.
Shuts down all processes and deletes all state associated with the
instance.
"""
if (not self._shutdown) and ray.is_initialized():
ray.get(self._controller.shutdown.remote())
... | def shutdown(self) -> None:
"""Completely shut down the connected Serve instance.
Shuts down all processes and deletes all state associated with the
instance.
"""
if not self._shutdown:
ray.get(self._controller.shutdown.remote())
ray.kill(self._contro... | [{'piece_type': 'source code', 'piece_content': 'from ray import serve\\n\\nclient = serve.start(detached=True)\\nclient.shutdown()\\nclient = serve.start(detached=True)'}, {'piece_type': 'error message', 'piece_content': 'File descriptor limit 256 is too low for production servers and may result in connection errors. ... | File descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with 'ulimit -n 8192'
2020-11-20 10:38:27,065 INFO services.py:1173 -- View the Ray dashboard at http://127.0.0.1:8265
(pid=raylet) 2020-11-20 10:38:29,628 INFO controller.py:313 -- Sta... | ray.serve.exceptions.RayServeException |
def __init__(
self,
obs_space: gym.spaces.Space,
action_space: gym.spaces.Space,
config: TrainerConfigDict,
loss_fn: Callable[[
Policy, ModelV2, Type[TFActionDistribution], SampleBatch
], TensorType],
*,
... | def __init__(
self,
obs_space: gym.spaces.Space,
action_space: gym.spaces.Space,
config: TrainerConfigDict,
loss_fn: Callable[[
Policy, ModelV2, Type[TFActionDistribution], SampleBatch
], TensorType],
*,
... | [{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi... | WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version.
Instructions for updating:
This function will only be available ... | tensorflow.python.ops.op_selector.UnliftableError |
def __init__(self,
observation_space: gym.spaces.Space,
action_space: gym.spaces.Space,
config: TrainerConfigDict,
sess: "tf1.Session",
obs_input: TensorType,
sampled_action: TensorType,
loss: Tens... | def __init__(self,
observation_space: gym.spaces.Space,
action_space: gym.spaces.Space,
config: TrainerConfigDict,
sess: "tf1.Session",
obs_input: TensorType,
sampled_action: TensorType,
loss: Tens... | [{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi... | WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version.
Instructions for updating:
This function will only be available ... | tensorflow.python.ops.op_selector.UnliftableError |
def try_import_tf(error=False):
"""Tries importing tf and returns the module (or None).
Args:
error (bool): Whether to raise an error if tf cannot be imported.
Returns:
Tuple:
- tf1.x module (either from tf2.x.compat.v1 OR as tf1.x).
- tf module (resulting from `imp... | def try_import_tf(error=False):
"""Tries importing tf and returns the module (or None).
Args:
error (bool): Whether to raise an error if tf cannot be imported.
Returns:
Tuple:
- tf1.x module (either from tf2.x.compat.v1 OR as tf1.x).
- tf module (resulting from `imp... | [{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi... | WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version.
Instructions for updating:
This function will only be available ... | tensorflow.python.ops.op_selector.UnliftableError |
def build_trainer(
name: str,
*,
default_config: Optional[TrainerConfigDict] = None,
validate_config: Optional[Callable[[TrainerConfigDict], None]] = None,
default_policy: Optional[Type[Policy]] = None,
get_policy_class: Optional[Callable[[TrainerConfigDict], Optional[Typ... | def build_trainer(
name: str,
*,
default_config: Optional[TrainerConfigDict] = None,
validate_config: Optional[Callable[[TrainerConfigDict], None]] = None,
default_policy: Optional[Type[Policy]] = None,
get_policy_class: Optional[Callable[[TrainerConfigDict], Optional[Typ... | [{'piece_type': 'other', 'piece_content': 'from or_gym.utils import create_env\\nfrom gym import spaces\\nfrom ray.rllib.utils import try_import_tf\\nfrom ray.rllib.models.tf.fcnet import FullyConnectedNetwork\\nfrom ray.rllib.models.tf.tf_modelv2 import TFModelV2\\nfrom ray.rllib.models import ModelCatalog\\nfrom ray ... | Traceback (most recent call last):
File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/ray/rllib/models/preprocessors.py", line 60, in check_shape
if not self._obs_space.contains(observation):
File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/gym/spaces/box.py", line 128, in co... | AttributeError |
def __init__(self,
*,
env_creator: Optional[Callable[[EnvContext], EnvType]] = None,
validate_env: Optional[Callable[[EnvType], None]] = None,
policy_class: Optional[Type[Policy]] = None,
trainer_config: Optional[TrainerConfigDict]... | def __init__(self,
*,
env_creator: Optional[Callable[[EnvContext], EnvType]] = None,
validate_env: Optional[Callable[[EnvType], None]] = None,
policy_class: Optional[Type[Policy]] = None,
trainer_config: Optional[TrainerConfigDict]... | [{'piece_type': 'other', 'piece_content': 'from or_gym.utils import create_env\\nfrom gym import spaces\\nfrom ray.rllib.utils import try_import_tf\\nfrom ray.rllib.models.tf.fcnet import FullyConnectedNetwork\\nfrom ray.rllib.models.tf.tf_modelv2 import TFModelV2\\nfrom ray.rllib.models import ModelCatalog\\nfrom ray ... | Traceback (most recent call last):
File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/ray/rllib/models/preprocessors.py", line 60, in check_shape
if not self._obs_space.contains(observation):
File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/gym/spaces/box.py", line 128, in co... | AttributeError |
def memory_summary():
"""Returns a formatted string describing memory usage in the cluster."""
import grpc
from ray.core.generated import node_manager_pb2
from ray.core.generated import node_manager_pb2_grpc
# We can ask any Raylet for the global memory info.
raylet = ray.nodes()[0]
raylet... | def memory_summary():
"""Returns a formatted string describing memory usage in the cluster."""
import grpc
from ray.core.generated import node_manager_pb2
from ray.core.generated import node_manager_pb2_grpc
# We can ask any Raylet for the global memory info.
raylet = ray.nodes()[0]
raylet... | [{'piece_type': 'error message', 'piece_content': '2020-05-19 02:13:32,283\\tINFO scripts.py:976 -- Connecting to Ray instance at 172.31.6.12:34940.\\n2020-05-19 02:13:32,284\\tWARNING worker.py:809 -- When connecting to an existing cluster, _internal_config must match the cluster\\'s _internal_config.\\n(pid=5906) E05... | 2020-05-19 02:13:32,283 INFO scripts.py:976 -- Connecting to Ray instance at 172.31.6.12:34940.
2020-05-19 02:13:32,284 WARNING worker.py:809 -- When connecting to an existing cluster, _internal_config must match the cluster's _internal_config.
(pid=5906) E0519 02:13:32.383447 5906 plasma_store_provider.cc:108] Failed... | grpc._channel._InactiveRpcError |
def custom_excepthook(type, value, tb):
# If this is a driver, push the exception to GCS worker table.
if global_worker.mode == SCRIPT_MODE:
error_message = "".join(traceback.format_tb(tb))
worker_id = global_worker.worker_id
worker_type = ray.gcs_utils.DRIVER
worker_info = {"exc... | def custom_excepthook(type, value, tb):
# If this is a driver, push the exception to GCS worker table.
if global_worker.mode == SCRIPT_MODE:
error_message = "".join(traceback.format_tb(tb))
worker_id = global_worker.worker_id
worker_type = ray.gcs_utils.DRIVER
worker_info = {"exc... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nray.init()\\n\\nraise Exception("foobar")'}, {'piece_type': 'error message', 'piece_content': 'raise Exception("hello")\\nError in sys.excepthook:\\nTraceback (most recent call last):\\nFile "/Users/eoakes/code/ray/python/ray/worker.py", line 836, in custom_... | raise Exception("hello")
Error in sys.excepthook:
Traceback (most recent call last):
File "/Users/eoakes/code/ray/python/ray/worker.py", line 836, in custom_excepthook
ray.state.state.add_worker(worker_id, worker_type, worker_info)
File "/Users/eoakes/code/ray/python/ray/state.py", line 733, in add_worker
return self.g... | AttributeError |
def __init__(self,
sync_up_template,
sync_down_template,
delete_template=noop_template):
"""Syncs between two directories with the given command.
Arguments:
sync_up_template (str): A runnable string template; needs to
in... | def __init__(self,
sync_up_template,
sync_down_template,
delete_template=noop_template):
"""Syncs between two directories with the given command.
Arguments:
sync_up_template (str): A runnable string template; needs to
in... | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
def set_logdir(self, logdir):
"""Sets the directory to log sync execution output in.
Args:
logdir (str): Log directory.
"""
self.logfile = tempfile.NamedTemporaryFile(
prefix="log_sync_out", dir=logdir, suffix=".log", delete=False)
self._closed = Fals... | def set_logdir(self, logdir):
"""Sets the directory to log sync execution output in.
Args:
logdir (str): Log directory.
"""
self.logfile = tempfile.NamedTemporaryFile(
prefix="log_sync_out", dir=logdir, suffix=".log", delete=False) | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
def delete(self, target):
if self.is_running:
logger.warning("Last sync client cmd still in progress, skipping.")
return False
final_cmd = self.delete_template.format(target=quote(target))
logger.debug("Running delete: {}".format(final_cmd))
self.cmd_process =... | def delete(self, target):
if self.is_running:
logger.warning("Last sync client cmd still in progress, skipping.")
return False
final_cmd = self.delete_template.format(target=quote(target))
logger.debug("Running delete: {}".format(final_cmd))
self.cmd_process =... | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
def _execute(self, sync_template, source, target):
"""Executes sync_template on source and target."""
if self.is_running:
logger.warning("Last sync client cmd still in progress, skipping.")
return False
final_cmd = sync_template.format(
source=quote(source... | def _execute(self, sync_template, source, target):
"""Executes sync_template on source and target."""
if self.is_running:
logger.warning("Last sync client cmd still in progress, skipping.")
return False
final_cmd = sync_template.format(
source=quote(source... | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
def on_trial_complete(self, iteration: int, trials: List["Trial"],
trial: "Trial", **info):
trial_syncer = self._get_trial_syncer(trial)
if NODE_IP in trial.last_result:
trainable_ip = trial.last_result[NODE_IP]
else:
trainable_ip = ray.get(t... | def on_trial_complete(self, iteration: int, trials: List["Trial"],
trial: "Trial", **info):
trial_syncer = self._get_trial_syncer(trial)
if NODE_IP in trial.last_result:
trainable_ip = trial.last_result[NODE_IP]
else:
trainable_ip = ray.get(t... | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
def mock_storage_client():
"""Mocks storage client that treats a local dir as durable storage."""
client = get_sync_client(LOCAL_SYNC_TEMPLATE, LOCAL_DELETE_TEMPLATE)
path = os.path.join(ray.utils.get_user_temp_dir(),
f"mock-client-{uuid.uuid4().hex[:4]}")
os.makedirs(path, exist... | def mock_storage_client():
"""Mocks storage client that treats a local dir as durable storage."""
return get_sync_client(LOCAL_SYNC_TEMPLATE, LOCAL_DELETE_TEMPLATE) | [{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt... | 2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down
self._local_dir)
File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package... | OSError |
async def _do_long_poll(self):
while True:
try:
updates: Dict[str, UpdatedObject] = await self._poll_once()
self._update(updates)
logger.debug(f"LongPollerClient received udpates: {updates}")
for key, updated_object in updates.items... | async def _do_long_poll(self):
while True:
updates: Dict[str, UpdatedObject] = await self._poll_once()
self._update(updates)
logger.debug(f"LongPollerClient received updates: {updates}")
for key, updated_object in updates.items():
# NOTE(simon)... | [{'piece_type': 'source code', 'piece_content': 'import ray\\nimport asyncio\\n\\nray.init()\\n@ray.remote\\nclass A:\\ndef hi(self): return "hi"\\n\\na = A.remote()\\nasync def main():\\nawait a.hi.remote()\\nray.kill(a)\\nawait a.hi.remote()\\n\\nasyncio.get_event_loop().run_until_complete(main())'}, {'piece_type': '... | Exception in callback async_set_result.<locals>.set_future()
handle: <Handle async_set_result.<locals>.set_future()>
Traceback (most recent call last):
File "/Users/simonmo/miniconda3/lib/python3.6/asyncio/events.py", line 145, in _run
self._callback(*self._args)
File "python/ray/_raylet.pyx", line 1530, in ray._raylet... | AttributeError |
async def _get_actor(actor):
actor = dict(actor)
worker_id = actor["address"]["workerId"]
core_worker_stats = DataSource.core_worker_stats.get(worker_id, {})
actor_constructor = core_worker_stats.get("actorTitle",
"Unknown actor const... | async def _get_actor(actor):
actor = dict(actor)
worker_id = actor["address"]["workerId"]
core_worker_stats = DataSource.core_worker_stats.get(worker_id, {})
actor_constructor = core_worker_stats.get("actorTitle",
"Unknown actor const... | [{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 347, in _update_cache response = task.result() File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_... | Error: Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 347, in _update_cache response = task.result() File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 77, in get... | TypeError |
def get_address_info_from_redis_helper(redis_address,
node_ip_address,
redis_password=None):
redis_ip_address, redis_port = redis_address.split(":")
# Get node table from global state accessor.
global_state = ray.state.GlobalState... | def get_address_info_from_redis_helper(redis_address,
node_ip_address,
redis_password=None):
redis_ip_address, redis_port = redis_address.split(":")
# Get node table from global state accessor.
global_state = ray.state.GlobalState... | [{'piece_type': 'error message', 'piece_content': '020-11-11 14:48:28,960\\tINFO worker.py:672 -- Connecting to existing Ray cluster at address: ***:***\\n2020-11-11 14:48:28,968\\tWARNING services.py:218 -- Some processes that the driver needs to connect to have not registered with Redis, so retrying. Have you run \\'... | 020-11-11 14:48:28,960 INFO worker.py:672 -- Connecting to existing Ray cluster at address: ***:***
2020-11-11 14:48:28,968 WARNING services.py:218 -- Some processes that the driver needs to connect to have not registered with Redis, so retrying. Have you run 'ray start' on this node?
2020-11-11 14:48:29,977 WARNING se... | RuntimeError |
def __init__(self):
self.indent_level = 0
self._verbosity = 0
self._verbosity_overriden = False
self._color_mode = "auto"
self._log_style = "record"
self.pretty = False
self.interactive = False
# store whatever colorful has detected for future use if... | def __init__(self):
self.indent_level = 0
self._verbosity = 0
self._color_mode = "auto"
self._log_style = "record"
self.pretty = False
self.interactive = False
# store whatever colorful has detected for future use if
# the color ouput is toggled (col... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def verbosity(self):
if self._verbosity_overriden:
return self._verbosity
elif not self.pretty:
return 999
return self._verbosity | def verbosity(self):
if not self.pretty:
return 999
return self._verbosity | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def _set_verbosity(self, x):
self._verbosity = x
self._verbosity_overriden = True | def _set_verbosity(self, x):
self._verbosity = x | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def rsync(config_file: str,
source: Optional[str],
target: Optional[str],
override_cluster_name: Optional[str],
down: bool,
ip_address: Optional[str] = None,
use_internal_ip: bool = False,
no_config_cache: bool = False,
all_nodes: bool = Fa... | def rsync(config_file: str,
source: Optional[str],
target: Optional[str],
override_cluster_name: Optional[str],
down: bool,
ip_address: Optional[str] = None,
use_internal_ip: bool = False,
no_config_cache: bool = False,
all_nodes: bool = Fa... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def rsync_to_node(node_id, is_head_node):
updater = NodeUpdaterThread(
node_id=node_id,
provider_config=config["provider"],
provider=provider,
auth_config=config["auth"],
cluster_name=config["cluster_name"],
file_mounts=config["file_mou... | def rsync_to_node(node_id, is_head_node):
updater = NodeUpdaterThread(
node_id=node_id,
provider_config=config["provider"],
provider=provider,
auth_config=config["auth"],
cluster_name=config["cluster_name"],
file_mounts=config["file_mou... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def __init__(self,
local_dir: str,
remote_dir: str,
sync_client: Optional[SyncClient] = None):
configure_logging(
log_style="record",
verbosity=env_integer("TUNE_SYNCER_VERBOSITY", 0))
self.local_ip = services.get_node_ip_add... | def __init__(self,
local_dir: str,
remote_dir: str,
sync_client: Optional[SyncClient] = None):
self.local_ip = services.get_node_ip_address()
self.worker_ip = None
sync_client = sync_client or DockerSyncClient()
sync_client.configur... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def reset_trial(self,
trial,
new_config,
new_experiment_tag,
logger_creator=None):
"""Tries to invoke `Trainable.reset()` to reset trial.
Args:
trial (Trial): Trial to be reset.
new_config (dict)... | def reset_trial(self,
trial,
new_config,
new_experiment_tag,
logger_creator=None):
"""Tries to invoke `Trainable.reset()` to reset trial.
Args:
trial (Trial): Trial to be reset.
new_config (dict)... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def reset(self, new_config, logger_creator=None):
"""Resets trial for use with new config.
Subclasses should override reset_config() to actually
reset actor behavior for the new config."""
self.config = new_config
self._result_logger.flush()
self._result_logger.clos... | def reset(self, new_config, logger_creator=None):
"""Resets trial for use with new config.
Subclasses should override reset_config() to actually
reset actor behavior for the new config."""
self.config = new_config
self._result_logger.flush()
self._result_logger.clos... | [{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64... | 2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn
ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local)
2020-11-19 19:42:27,247... | ray.tune.error.TuneError |
def _bootstrap_config(config: Dict[str, Any],
no_config_cache: bool = False) -> Dict[str, Any]:
config = prepare_config(config)
hasher = hashlib.sha1()
hasher.update(json.dumps([config], sort_keys=True).encode("utf-8"))
cache_key = os.path.join(tempfile.gettempdir(),
... | def _bootstrap_config(config: Dict[str, Any],
no_config_cache: bool = False) -> Dict[str, Any]:
config = prepare_config(config)
hasher = hashlib.sha1()
hasher.update(json.dumps([config], sort_keys=True).encode("utf-8"))
cache_key = os.path.join(tempfile.gettempdir(),
... | [{'piece_type': 'other', 'piece_content': '(base) ➜ tune git:(fix-kubernetes-dep) ✗ pip list | grep boto\\nboto 2.49.0\\nboto3 1.4.8\\nbotocore 1.8.50'}, {'piece_type': 'error message', 'piece_content': '(base) ➜ tune git:(fix-kubernetes-dep) ✗ ray... | (base) ➜ tune git:(fix-kubernetes-dep) ✗ ray up $CFG -y
Cluster: basic
Checking AWS environment settings
Traceback (most recent call last):
File "/Users/rliaw/miniconda3/bin/ray", line 8, in <module>
sys.exit(main())
File "/Users/rliaw/miniconda3/lib/python3.7/site-packages/ray/scripts/scripts.py", line 1471, in main... | AttributeError |
async def get_node_workers(cls, node_id):
workers = []
node_ip = DataSource.node_id_to_ip[node_id]
node_logs = DataSource.ip_and_pid_to_logs.get(node_ip, {})
node_errs = DataSource.ip_and_pid_to_errors.get(node_ip, {})
node_physical_stats = DataSource.node_physical_stats.get(... | async def get_node_workers(cls, node_id):
workers = []
node_ip = DataSource.node_id_to_ip[node_id]
node_logs = DataSource.ip_and_pid_to_logs.get(node_ip, {})
logger.error(node_logs)
node_errs = DataSource.ip_and_pid_to_errors.get(node_ip, {})
logger.error(node_errs)
... | [{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()\\nFile "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/st... | Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()
File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 89, in get_all_nodes all... | KeyError |
async def get_node_info(cls, node_id):
node_physical_stats = dict(
DataSource.node_physical_stats.get(node_id, {}))
node_stats = dict(DataSource.node_stats.get(node_id, {}))
node = DataSource.nodes.get(node_id, {})
node_ip = DataSource.node_id_to_ip.get(node_id)
#... | async def get_node_info(cls, node_id):
node_physical_stats = dict(
DataSource.node_physical_stats.get(node_id, {}))
node_stats = dict(DataSource.node_stats.get(node_id, {}))
node = DataSource.nodes.get(node_id, {})
node_ip = DataSource.node_id_to_ip.get(node_id)
#... | [{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()\\nFile "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/st... | Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()
File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 89, in get_all_nodes all... | KeyError |
def unflatten_dict(dt, delimiter="/"):
"""Unflatten dict. Does not support unflattening lists."""
dict_type = type(dt)
out = dict_type()
for key, val in dt.items():
path = key.split(delimiter)
item = out
for k in path[:-1]:
item = item.setdefault(k, dict_type())
... | def unflatten_dict(dt, delimiter="/"):
"""Unflatten dict. Does not support unflattening lists."""
out = defaultdict(dict)
for key, val in dt.items():
path = key.split(delimiter)
item = out
for k in path[:-1]:
item = item[k]
item[path[-1]] = val
return dict(out... | [{'piece_type': 'error message', 'piece_content': '$ python ./python/ray/tune/examples/bohb_example.py\\nFile descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with \\'ulimit -n 8192\\'\\n2020-11-11 20:38:44,944 INFO services.py:1110 -- Vie... | $ python ./python/ray/tune/examples/bohb_example.py
File descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with 'ulimit -n 8192'
2020-11-11 20:38:44,944 INFO services.py:1110 -- View the Ray dashboard at http://127.0.0.1:8265
Traceback (mos... | KeyError |
def find_redis_address(address=None):
pids = psutil.pids()
redis_addresses = set()
for pid in pids:
try:
proc = psutil.Process(pid)
# HACK: Workaround for UNIX idiosyncrasy
# Normally, cmdline() is supposed to return the argument list.
# But it in some... | def find_redis_address(address=None):
pids = psutil.pids()
redis_addresses = set()
for pid in pids:
try:
proc = psutil.Process(pid)
# HACK: Workaround for UNIX idiosyncrasy
# Normally, cmdline() is supposed to return the argument list.
# But it in some... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/swang/anaconda3/envs/ray-36/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1462, in main\\nreturn cli()\\nFile "/home/... | Traceback (most recent call last):
File "/home/swang/anaconda3/envs/ray-36/bin/ray", line 8, in <module>
sys.exit(main())
File "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1462, in main
return cli()
File "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/click... | ConnectionError |
def _run_helper(self,
final_cmd,
with_output=False,
exit_on_fail=False,
silent=False):
"""Run a command that was already setup with SSH and `bash` settings.
Args:
cmd (List[str]):
Full comman... | def _run_helper(self,
final_cmd,
with_output=False,
exit_on_fail=False,
silent=False):
"""Run a command that was already setup with SSH and `bash` settings.
Args:
cmd (List[str]):
Full comman... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "XXX/lib/python3.7/site-packages/ray/autoscaler/command_runner.py", line 248, in run\\nself.process_runner.check_call(final_cmd, shell=True)\\nFile "/Users/mkoh/.pyenv/versions/3.7.7/lib/python3.7/subprocess.py", line 363, in c... | Traceback (most recent call last):
File "XXX/lib/python3.7/site-packages/ray/autoscaler/command_runner.py", line 248, in run
self.process_runner.check_call(final_cmd, shell=True)
File "/Users/mkoh/.pyenv/versions/3.7.7/lib/python3.7/subprocess.py", line 363, in check_call
raise CalledProcessError(retcode, cmd)
subproce... | subprocess.CalledProcessError |
def __init__(self,
space: Optional[Union[Dict, List[Dict]]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
parameter_constraints: Optional[List] = None,
outcome_constraints: Optional[List] = None,
... | def __init__(self,
space: Optional[List[Dict]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
parameter_constraints: Optional[List] = None,
outcome_constraints: Optional[List] = None,
ax_client: ... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
space: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
utility_kwargs: Optional[Dict] = None,
random_state: int = 42,
random_search_steps: int = 10,
... | def __init__(self,
space: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
utility_kwargs: Optional[Dict] = None,
random_state: int = 42,
random_search_steps: int = 10,
... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def convert_search_space(spec: Dict, join: bool = False) -> Dict:
spec = flatten_dict(spec, prevent_delimiter=True)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if grid_vars:
raise ValueError(
"Grid search parameters cannot be automatically conve... | def convert_search_space(spec: Dict) -> Dict:
spec = flatten_dict(spec, prevent_delimiter=True)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if grid_vars:
raise ValueError(
"Grid search parameters cannot be automatically converted "
... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
space: Optional[Union[Dict,
ConfigSpace.ConfigurationSpace]] = None,
bohb_config: Optional[Dict] = None,
max_concurrent: int = 10,
metric: Optional[str] = None,
mode: Option... | def __init__(self,
space: Optional[ConfigSpace.ConfigurationSpace] = None,
bohb_config: Optional[Dict] = None,
max_concurrent: int = 10,
metric: Optional[str] = None,
mode: Optional[str] = None):
from hpbandster.optimizers.... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
optimizer: Optional[BlackboxOptimiser] = None,
domain: Optional[str] = None,
space: Optional[Union[Dict, List[Dict]]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_e... | def __init__(self,
optimizer: Optional[BlackboxOptimiser] = None,
domain: Optional[str] = None,
space: Optional[List[Dict]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_evaluate: Opti... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(
self,
space: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_evaluate: Optional[List[Dict]] = None,
n_initial_points: int = 20,
random_state_seed: Optional[int] = None,
... | def __init__(
self,
space: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_evaluate: Optional[List[Dict]] = None,
n_initial_points: int = 20,
random_state_seed: Optional[int] = None,
... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
optimizer: Union[None, Optimizer, ConfiguredOptimizer] = None,
space: Optional[Union[Dict, Parameter]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
max_concurrent: Optional[int] = None,
... | def __init__(self,
optimizer: Union[None, Optimizer, ConfiguredOptimizer] = None,
space: Optional[Parameter] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
max_concurrent: Optional[int] = None,
*... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
space: Optional[Union[Dict, List[Tuple]]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
sampler: Optional[BaseSampler] = None):
assert ot is not None, (
"Optuna must be installed! Run `pip ... | def __init__(self,
space: Optional[List[Tuple]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
sampler: Optional[BaseSampler] = None):
assert ot is not None, (
"Optuna must be installed! Run `pip install optun... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
optimizer: Optional[sko.optimizer.Optimizer] = None,
space: Union[List[str], Dict[str, Union[Tuple, List]]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_evaluate: Optional[List[List... | def __init__(self,
optimizer: Optional[sko.optimizer.Optimizer] = None,
space: Union[List[str], Dict[str, Union[Tuple, List]]] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
points_to_evaluate: Optional[List[List... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def convert_search_space(spec: Dict, join: bool = False) -> Dict:
spec = flatten_dict(spec, prevent_delimiter=True)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if grid_vars:
raise ValueError(
"Grid search parameters cannot be automatically conve... | def convert_search_space(spec: Dict) -> Dict:
spec = flatten_dict(spec, prevent_delimiter=True)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if grid_vars:
raise ValueError(
"Grid search parameters cannot be automatically converted "
... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def __init__(self,
algo: str = "asracos",
budget: Optional[int] = None,
dim_dict: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
**kwargs):
assert zoopt is not None, "ZOOpt... | def __init__(self,
algo: str = "asracos",
budget: Optional[int] = None,
dim_dict: Optional[Dict] = None,
metric: Optional[str] = None,
mode: Optional[str] = None,
**kwargs):
assert zoopt is not None, "ZOOpt... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def convert_search_space(spec: Dict,
join: bool = False) -> Dict[str, Tuple]:
spec = copy.deepcopy(spec)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if not domain_vars and not grid_vars:
return {}
if grid_vars:
... | def convert_search_space(spec: Dict) -> Dict[str, Tuple]:
spec = copy.deepcopy(spec)
resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec)
if not domain_vars and not grid_vars:
return []
if grid_vars:
raise ValueError(
"Grid search pa... | [{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '... | Traceback (most recent call last):
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner
(pid=82997) self.run()
(pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run
(pid=82997) ... | TypeError |
def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | def init(
address=None,
*,
num_cpus=None,
num_gpus=None,
resources=None,
object_store_memory=None,
local_mode=False,
ignore_reinit_error=False,
include_dashboard=None,
dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP,
dashboard_por... | [{'piece_type': 'other', 'piece_content': "(pid=136534) F1009 07:02:26.344231 136534 136534 service_based_gcs_client.cc:207] Couldn't reconnect to GCS server. The last attempted GCS server address was 192.168.1.98:39953\\n(pid=136534) *** Check failure stack trace: ***\\n(pid=136534) @ 0x7f56afdb7f5d google::L... | ^CTraceback (most recent call last):
File "test.py", line 72, in <module>
**tune_kwargs)
File "/home/demattia/miniconda3/envs/test_tune/lib/python3.7/site-packages/ray/tune/tune.py", line 405, in run
runner.step()
File "/home/demattia/miniconda3/envs/test_tune/lib/python3.7/site-packages/ray/tune/trial_runner.py", line... | CError |
def with_parameters(fn, **kwargs):
"""Wrapper for function trainables to pass arbitrary large data objects.
This wrapper function will store all passed parameters in the Ray
object store and retrieve them when calling the function. It can thus
be used to pass arbitrary data, even datasets, to Tune trai... | def with_parameters(fn, **kwargs):
"""Wrapper for function trainables to pass arbitrary large data objects.
This wrapper function will store all passed parameters in the Ray
object store and retrieve them when calling the function. It can thus
be used to pass arbitrary data, even datasets, to Tune trai... | [{'piece_type': 'error message', 'piece_content': 'Failure # 1 (occurred at 2020-09-26_16-50-01)\\nTraceback (most recent call last):\\nFile "/home/karol/PycharmProjects/ray/python/ray/tune/trial_runner.py", line 518, in _process_trial\\nresult = self.trial_executor.fetch_result(trial)\\nFile "/home/karol/PycharmProjec... | Failure # 1 (occurred at 2020-09-26_16-50-01)
Traceback (most recent call last):
File "/home/karol/PycharmProjects/ray/python/ray/tune/trial_runner.py", line 518, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "/home/karol/PycharmProjects/ray/python/ray/tune/ray_trial_executor.py", line 488, in... | ray.tune.error.TuneError |
def wandb_mixin(func: Callable):
"""wandb_mixin
Weights and biases (https://www.wandb.com/) is a tool for experiment
tracking, model optimization, and dataset versioning. This Ray Tune
Trainable mixin helps initializing the Wandb API for use with the
``Trainable`` class or with `@wandb_mixin` for t... | def wandb_mixin(func: Callable):
"""wandb_mixin
Weights and biases (https://www.wandb.com/) is a tool for experiment
tracking, model optimization, and dataset versioning. This Ray Tune
Trainable mixin helps initializing the Wandb API for use with the
``Trainable`` class or with `@wandb_mixin` for t... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ubuntu/run_ray_tune.py", line 222, in <module>\\ntune_helsinki_(args)\\nFile "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_\\nray_wandb_func = wandb_mixin(ray_func)\\nFile "/home/ubuntu/anaconda3/lib/python3... | Traceback (most recent call last):
File "/home/ubuntu/run_ray_tune.py", line 222, in <module>
tune_helsinki_(args)
File "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_
ray_wandb_func = wandb_mixin(ray_func)
File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/tune/integration/wandb.py", line 142, in... | AttributeError |
def _init(self):
config = self.config.copy()
config.pop("callbacks", None) # Remove callbacks
try:
if config.get("logger_config", {}).get("wandb"):
logger_config = config.pop("logger_config")
wandb_config = logger_config.get("wandb").copy()
... | def _init(self):
config = self.config.copy()
config.pop("callbacks", None) # Remove callbacks
try:
if config.get("logger_config", {}).get("wandb"):
logger_config = config.pop("logger_config")
wandb_config = logger_config.get("wandb").copy()
... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ubuntu/run_ray_tune.py", line 222, in <module>\\ntune_helsinki_(args)\\nFile "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_\\nray_wandb_func = wandb_mixin(ray_func)\\nFile "/home/ubuntu/anaconda3/lib/python3... | Traceback (most recent call last):
File "/home/ubuntu/run_ray_tune.py", line 222, in <module>
tune_helsinki_(args)
File "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_
ray_wandb_func = wandb_mixin(ray_func)
File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/tune/integration/wandb.py", line 142, in... | AttributeError |
def Concurrently(ops: List[LocalIterator],
*,
mode="round_robin",
output_indexes=None,
round_robin_weights=None):
"""Operator that runs the given parent iterators concurrently.
Args:
mode (str): One of 'round_robin', 'async'. In 'round... | def Concurrently(ops: List[LocalIterator],
*,
mode="round_robin",
output_indexes=None,
round_robin_weights=None):
"""Operator that runs the given parent iterators concurrently.
Args:
mode (str): One of {'round_robin', 'async'}.
... | [{'piece_type': 'other', 'piece_content': 'from ray.tune.logger import pretty_print\\nfrom ray.rllib.agents.dqn.dqn import DEFAULT_CONFIG\\nfrom ray.rllib.agents.dqn.dqn import DQNTrainer\\nimport numpy as np\\n\\nconfig = DEFAULT_CONFIG.copy()\\nconfig["env"] = "CartPole-v0"\\nconfig["input"] = {"demo-out": 0.3, "samp... | Traceback (most recent call last):
File "/home/enes/ws/code/arl/mt/test/consume_experiences.py", line 16, in <module>
results = trainer.train()
File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 522, in train
raise e
File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray... | KeyError |
def ParallelRollouts(workers: WorkerSet, *, mode="bulk_sync",
num_async=1) -> LocalIterator[SampleBatch]:
"""Operator to collect experiences in parallel from rollout workers.
If there are no remote workers, experiences will be collected serially from
the local worker instance instead.
... | def ParallelRollouts(workers: WorkerSet, *, mode="bulk_sync",
num_async=1) -> LocalIterator[SampleBatch]:
"""Operator to collect experiences in parallel from rollout workers.
If there are no remote workers, experiences will be collected serially from
the local worker instance instead.
... | [{'piece_type': 'other', 'piece_content': 'from ray.tune.logger import pretty_print\\nfrom ray.rllib.agents.dqn.dqn import DEFAULT_CONFIG\\nfrom ray.rllib.agents.dqn.dqn import DQNTrainer\\nimport numpy as np\\n\\nconfig = DEFAULT_CONFIG.copy()\\nconfig["env"] = "CartPole-v0"\\nconfig["input"] = {"demo-out": 0.3, "samp... | Traceback (most recent call last):
File "/home/enes/ws/code/arl/mt/test/consume_experiences.py", line 16, in <module>
results = trainer.train()
File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 522, in train
raise e
File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray... | KeyError |
def _get_node_specific_docker_config(self, node_id):
if "docker" not in self.config:
return {}
docker_config = copy.deepcopy(self.config.get("docker", {}))
node_specific_docker = self._get_node_type_specific_fields(
node_id, "docker")
docker_config.update(node... | def _get_node_specific_docker_config(self, node_id):
docker_config = copy.deepcopy(self.config.get("docker", {}))
node_specific_docker = self._get_node_type_specific_fields(
node_id, "docker")
docker_config.update(node_specific_docker)
return docker_config | [{'piece_type': 'other', 'piece_content': 'min_workers: 1\\nmax_workers: 1\\ninitial_workers: 1'}, {'piece_type': 'other', 'piece_content': '(myenv) tcw@pop-os:~/projects/ray$ ray up example-full.yaml\\nCommands running under a login shell can produce more output than special processing can handle.\\nThus, the output f... | ==> /tmp/ray/session_2020-09-09_17-20-39_779593_74/logs/monitor.err <==
docker_config = self._get_node_specific_docker_config(node_id)
File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config
node_id, "docker")
File "/root/anaconda3/lib/python3.7/sit... | KeyError |
def _clean_log(obj):
# Fixes https://github.com/ray-project/ray/issues/10631
if isinstance(obj, dict):
return {k: _clean_log(v) for k, v in obj.items()}
elif isinstance(obj, list):
return [_clean_log(v) for v in obj]
# Else
try:
pickle.dumps(obj)
yaml.dump(
... | def _clean_log(obj):
# Fixes https://github.com/ray-project/ray/issues/10631
if isinstance(obj, dict):
return {k: _clean_log(v) for k, v in obj.items()}
elif isinstance(obj, list):
return [_clean_log(v) for v in obj]
# Else
try:
pickle.dumps(obj)
return obj
excep... | [{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi... | Process _WandbLoggingProcess-1:
Traceback (most recent call last):
File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap
self.run()
File "[...]/ray/tune/integration/wandb.py", line 127, in run
wandb.init(*self.args, **self.kwargs)
File "[...]/wandb/__init__.py", line 1303, in init
as_defaults=no... | yaml.representer.RepresenterError |
def _init(self):
config = self.config.copy()
config.pop("callbacks", None) # Remove callbacks
try:
if config.get("logger_config", {}).get("wandb"):
logger_config = config.pop("logger_config")
wandb_config = logger_config.get("wandb").copy()
... | def _init(self):
config = self.config.copy()
try:
if config.get("logger_config", {}).get("wandb"):
logger_config = config.pop("logger_config")
wandb_config = logger_config.get("wandb").copy()
else:
wandb_config = config.pop("wa... | [{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi... | Process _WandbLoggingProcess-1:
Traceback (most recent call last):
File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap
self.run()
File "[...]/ray/tune/integration/wandb.py", line 127, in run
wandb.init(*self.args, **self.kwargs)
File "[...]/wandb/__init__.py", line 1303, in init
as_defaults=no... | yaml.representer.RepresenterError |
def _clean_log(obj):
# Fixes https://github.com/ray-project/ray/issues/10631
if isinstance(obj, dict):
return {k: _clean_log(v) for k, v in obj.items()}
elif isinstance(obj, list):
return [_clean_log(v) for v in obj]
elif _is_allowed_type(obj):
return obj
# Else
try:
... | def _clean_log(obj):
# Fixes https://github.com/ray-project/ray/issues/10631
if isinstance(obj, dict):
return {k: _clean_log(v) for k, v in obj.items()}
elif isinstance(obj, list):
return [_clean_log(v) for v in obj]
# Else
try:
pickle.dumps(obj)
yaml.dump(
... | [{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi... | Process _WandbLoggingProcess-1:
Traceback (most recent call last):
File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap
self.run()
File "[...]/ray/tune/integration/wandb.py", line 127, in run
wandb.init(*self.args, **self.kwargs)
File "[...]/wandb/__init__.py", line 1303, in init
as_defaults=no... | yaml.representer.RepresenterError |
def _handle_result(self, result):
config_update = result.get("config", {}).copy()
log = {}
flat_result = flatten_dict(result, delimiter="/")
for k, v in flat_result.items():
if any(
k.startswith(item + "/") or k == item
for item in... | def _handle_result(self, result):
config_update = result.get("config", {}).copy()
log = {}
flat_result = flatten_dict(result, delimiter="/")
for k, v in flat_result.items():
if any(
k.startswith(item + "/") or k == item
for item in... | [{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi... | Process _WandbLoggingProcess-1:
Traceback (most recent call last):
File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap
self.run()
File "[...]/ray/tune/integration/wandb.py", line 127, in run
wandb.init(*self.args, **self.kwargs)
File "[...]/wandb/__init__.py", line 1303, in init
as_defaults=no... | yaml.representer.RepresenterError |
def start(node_ip_address, redis_address, address, redis_port, port,
num_redis_shards, redis_max_clients, redis_password,
redis_shard_ports, object_manager_port, node_manager_port,
gcs_server_port, min_worker_port, max_worker_port, memory,
object_store_memory, redis_max_memory, n... | def start(node_ip_address, redis_address, address, redis_port, port,
num_redis_shards, redis_max_clients, redis_password,
redis_shard_ports, object_manager_port, node_manager_port,
gcs_server_port, min_worker_port, max_worker_port, memory,
object_store_memory, redis_max_memory, n... | [{'piece_type': 'error message', 'piece_content': '$ ray memory\\n2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ubuntu/anaconda3/lib/pytho... | $ ray memory
2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>
sys.exit(main())
File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1602, in mai... | TypeError |
def memory(address, redis_password):
"""Print object references held in a Ray cluster."""
if not address:
address = services.find_redis_address_or_die()
logger.info(f"Connecting to Ray instance at {address}.")
ray.init(address=address, _redis_password=redis_password)
print(ray.internal.inter... | def memory(address, redis_password):
"""Print object references held in a Ray cluster."""
if not address:
address = services.find_redis_address_or_die()
logger.info(f"Connecting to Ray instance at {address}.")
ray.init(address=address, redis_password=redis_password)
print(ray.internal.intern... | [{'piece_type': 'error message', 'piece_content': '$ ray memory\\n2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ubuntu/anaconda3/lib/pytho... | $ ray memory
2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.
Traceback (most recent call last):
File "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>
sys.exit(main())
File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1602, in mai... | TypeError |
def choose_trial_to_run(self, trial_runner, allow_recurse=True):
"""Fair scheduling within iteration by completion percentage.
List of trials not used since all trials are tracked as state
of scheduler. If iteration is occupied (ie, no trials to run),
then look into next iteration.
... | def choose_trial_to_run(self, trial_runner):
"""Fair scheduling within iteration by completion percentage.
List of trials not used since all trials are tracked as state
of scheduler. If iteration is occupied (ie, no trials to run),
then look into next iteration.
"""
... | [{'piece_type': 'error message', 'piece_content': '== Status ==\\nMemory usage on this node: 7.0/15.6 GiB\\nUsing HyperBand: num_stopped=832 total_brackets=3\\nRound #0:\\nNone\\nBracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}\\nBracket(Max Size (n)=324, Milestone (r)=8, com... | == Status ==
Memory usage on this node: 7.0/15.6 GiB
Using HyperBand: num_stopped=832 total_brackets=3
Round #0:
None
Bracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}
Bracket(Max Size (n)=324, Milestone (r)=8, completed=47.3%): {PAUSED: 166}
Resources requested: 4/32 CPUs, 0/... | ray.tune.error.TuneError |
def debug_string(self):
"""This provides a progress notification for the algorithm.
For each bracket, the algorithm will output a string as follows:
Bracket(Max Size (n)=5, Milestone (r)=33, completed=14.6%):
{PENDING: 2, RUNNING: 3, TERMINATED: 2}
"Max Size" indic... | def debug_string(self):
"""This provides a progress notification for the algorithm.
For each bracket, the algorithm will output a string as follows:
Bracket(Max Size (n)=5, Milestone (r)=33, completed=14.6%):
{PENDING: 2, RUNNING: 3, TERMINATED: 2}
"Max Size" indic... | [{'piece_type': 'error message', 'piece_content': '== Status ==\\nMemory usage on this node: 7.0/15.6 GiB\\nUsing HyperBand: num_stopped=832 total_brackets=3\\nRound #0:\\nNone\\nBracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}\\nBracket(Max Size (n)=324, Milestone (r)=8, com... | == Status ==
Memory usage on this node: 7.0/15.6 GiB
Using HyperBand: num_stopped=832 total_brackets=3
Round #0:
None
Bracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}
Bracket(Max Size (n)=324, Milestone (r)=8, completed=47.3%): {PAUSED: 166}
Resources requested: 4/32 CPUs, 0/... | ray.tune.error.TuneError |
def run_rsync_up(self, source, target):
if target.startswith("~"):
target = "/root" + target[1:]
try:
self.process_runner.check_call([
KUBECTL_RSYNC,
"-avz",
source,
"{}@{}:{}".format(self.node_id, self.namespac... | def run_rsync_up(self, source, target):
if target.startswith("~"):
target = "/root" + target[1:]
try:
self.process_runner.check_call([
KUBECTL_RSYNC,
"-avz",
source,
"{}@{}:{}".format(self.node_id, self.namespac... | [{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call... | 2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/ra... | ray.tune.error.TuneError |
def run_rsync_down(self, source, target):
if target.startswith("~"):
target = "/root" + target[1:]
try:
self.process_runner.check_call([
KUBECTL_RSYNC,
"-avz",
"{}@{}:{}".format(self.node_id, self.namespace, source),
... | def run_rsync_down(self, source, target):
if target.startswith("~"):
target = "/root" + target[1:]
try:
self.process_runner.check_call([
KUBECTL_RSYNC,
"-avz",
"{}@{}:{}".format(self.node_id, self.namespace, source),
... | [{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call... | 2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/ra... | ray.tune.error.TuneError |
def get_node_syncer(local_dir, remote_dir=None, sync_function=None):
"""Returns a NodeSyncer.
Args:
local_dir (str): Source directory for syncing.
remote_dir (str): Target directory for syncing. If not provided, a
noop Syncer is returned.
sync_function (func|str|bool): Funct... | def get_node_syncer(local_dir, remote_dir=None, sync_function=None):
"""Returns a NodeSyncer.
Args:
local_dir (str): Source directory for syncing.
remote_dir (str): Target directory for syncing. If not provided, a
noop Syncer is returned.
sync_function (func|str|bool): Funct... | [{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call... | 2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/ra... | ray.tune.error.TuneError |
def action_prob(self, batch: SampleBatchType) -> np.ndarray:
"""Returns the probs for the batch actions for the current policy."""
num_state_inputs = 0
for k in batch.keys():
if k.startswith("state_in_"):
num_state_inputs += 1
state_keys = ["state_in_{}".... | def action_prob(self, batch: SampleBatchType) -> TensorType:
"""Returns the probs for the batch actions for the current policy."""
num_state_inputs = 0
for k in batch.keys():
if k.startswith("state_in_"):
num_state_inputs += 1
state_keys = ["state_in_{}".... | [{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "C:\\\\Users\\\\Julius\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\tune\\\\trial_runner.py", line 497, in _process_trial\\nresult = self.trial_executor.fetch_result(trial)\\nFile "C:\\\\Users\\\\Julius\\\\Anac... | Traceback (most recent call last):
File "C:\\Users\\Julius\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\tune\\trial_runner.py", line 497, in _process_trial
result = self.trial_executor.fetch_result(trial)
File "C:\\Users\\Julius\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\tune\\ray_trial_executor.py", line 434, ... | AttributeError |
def run_rsync_up(self, source, target):
# TODO(ilr) Expose this to before NodeUpdater::sync_file_mounts
protected_path = target
if target.find("/root") == 0:
target = target.replace("/root", "/tmp/root")
self.ssh_command_runner.run(
f"mkdir -p {os.path.dirname... | def run_rsync_up(self, source, target):
protected_path = target
if target.find("/root") == 0:
target = target.replace("/root", "/tmp/root")
self.ssh_command_runner.run(
f"mkdir -p {os.path.dirname(target.rstrip('/'))}")
self.ssh_command_runner.run_rsync_up(sou... | [{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p... | (vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml
2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1
2020-08-12 20:12:39,612 INFO c... | AssertionError |
def sync_file_mounts(self, sync_cmd, step_numbers=(0, 2)):
# step_numbers is (# of previous steps, total steps)
previous_steps, total_steps = step_numbers
nolog_paths = []
if cli_logger.verbosity == 0:
nolog_paths = [
"~/ray_bootstrap_key.pem", "~/ray_boo... | def sync_file_mounts(self, sync_cmd, step_numbers=(0, 2)):
# step_numbers is (# of previous steps, total steps)
previous_steps, total_steps = step_numbers
nolog_paths = []
if cli_logger.verbosity == 0:
nolog_paths = [
"~/ray_bootstrap_key.pem", "~/ray_boo... | [{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p... | (vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml
2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1
2020-08-12 20:12:39,612 INFO c... | AssertionError |
def do_sync(remote_path, local_path, allow_non_existing_paths=False):
if allow_non_existing_paths and not os.path.exists(local_path):
# Ignore missing source files. In the future we should support
# the --delete-missing-args command to delete files that have
... | def do_sync(remote_path, local_path, allow_non_existing_paths=False):
if allow_non_existing_paths and not os.path.exists(local_path):
# Ignore missing source files. In the future we should support
# the --delete-missing-args command to delete files that have
... | [{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p... | (vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml
2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1
2020-08-12 20:12:39,612 INFO c... | AssertionError |
def wait_ready(self, deadline):
with cli_logger.group(
"Waiting for SSH to become available", _numbered=("[]", 1, 6)):
with LogTimer(self.log_prefix + "Got remote shell"):
cli_logger.old_info(logger, "{}Waiting for remote shell...",
... | def wait_ready(self, deadline):
with cli_logger.group(
"Waiting for SSH to become available", _numbered=("[]", 1, 6)):
with LogTimer(self.log_prefix + "Got remote shell"):
cli_logger.old_info(logger, "{}Waiting for remote shell...",
... | [{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p... | (vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml
2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1
2020-08-12 20:12:39,612 INFO c... | AssertionError |
def create_or_update_cluster(config_file: str,
override_min_workers: Optional[int],
override_max_workers: Optional[int],
no_restart: bool,
restart_only: bool,
yes: bool,
... | def create_or_update_cluster(
config_file: str, override_min_workers: Optional[int],
override_max_workers: Optional[int], no_restart: bool,
restart_only: bool, yes: bool, override_cluster_name: Optional[str],
no_config_cache: bool, dump_command_output: bool,
use_login_shells: boo... | [{'piece_type': 'error message', 'piece_content': '(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py\\nTraceback (most recent call last):\\nFile "/Users/alex/miniconda3/bin/ray", line 11, in <module>\\nload_entry_point(\\'ray\\', \\'console_scripts\\', \\'ray\\')()\\nFile "/Users/alex/anyscale/... | (base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py
Traceback (most recent call last):
File "/Users/alex/miniconda3/bin/ray", line 11, in <module>
load_entry_point('ray', 'console_scripts', 'ray')()
File "/Users/alex/anyscale/ray/python/ray/scripts/scripts.py", line 1587, in main
return cli()
Fil... | TypeError |
def submit(cluster_config_file, screen, tmux, stop, start, cluster_name,
port_forward, script, args, script_args, log_new_style, log_color,
verbose):
"""Uploads and runs a script on the specified cluster.
The script is automatically synced to the following location:
os.path.join(... | def submit(cluster_config_file, screen, tmux, stop, start, cluster_name,
port_forward, script, args, script_args, log_new_style, log_color,
verbose):
"""Uploads and runs a script on the specified cluster.
The script is automatically synced to the following location:
os.path.join(... | [{'piece_type': 'error message', 'piece_content': '(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py\\nTraceback (most recent call last):\\nFile "/Users/alex/miniconda3/bin/ray", line 11, in <module>\\nload_entry_point(\\'ray\\', \\'console_scripts\\', \\'ray\\')()\\nFile "/Users/alex/anyscale/... | (base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py
Traceback (most recent call last):
File "/Users/alex/miniconda3/bin/ray", line 11, in <module>
load_entry_point('ray', 'console_scripts', 'ray')()
File "/Users/alex/anyscale/ray/python/ray/scripts/scripts.py", line 1587, in main
return cli()
Fil... | TypeError |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.