after_merge
stringlengths
64
17k
before_merge
stringlengths
60
17k
source code and errors
stringlengths
236
32.3k
full_traceback
stringlengths
170
17.7k
traceback_type
stringclasses
60 values
def _blocking_send(self, req: ray_client_pb2.DataRequest ) -> ray_client_pb2.DataResponse: req_id = self._next_id() req.req_id = req_id self.request_queue.put(req) data = None with self.cv: self.cv.wait_for( lambda: req_id in...
def _blocking_send(self, req: ray_client_pb2.DataRequest ) -> ray_client_pb2.DataResponse: req_id = self._next_id() req.req_id = req_id self.request_queue.put(req) data = None with self.cv: self.cv.wait_for(lambda: req_id in self.ready_data)...
[{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive...
Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "Received http2 header with status: 502" debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr...
debug_error
def _log_main(self) -> None: stub = ray_client_pb2_grpc.RayletLogStreamerStub(self.channel) log_stream = stub.Logstream( iter(self.request_queue.get, None), metadata=self._metadata) try: for record in log_stream: if record.level < 0: ...
def _log_main(self) -> None: stub = ray_client_pb2_grpc.RayletLogStreamerStub(self.channel) log_stream = stub.Logstream( iter(self.request_queue.get, None), metadata=self._metadata) try: for record in log_stream: if record.level < 0: ...
[{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive...
Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "Received http2 header with status: 502" debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr...
debug_error
def __init__(self, conn_str: str = "", secure: bool = False, metadata: List[Tuple[str, str]] = None, connection_retries: int = 3): """Initializes the worker side grpc client. Args: conn_str: The host:port connection str...
def __init__(self, conn_str: str = "", secure: bool = False, metadata: List[Tuple[str, str]] = None, connection_retries: int = 3): """Initializes the worker side grpc client. Args: conn_str: The host:port connection str...
[{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610421939.184399675","description":"Receive...
Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "Received http2 header with status: 502" debug_error_string = "{"created":"@1610421939.184399675","description":"Received http2 :status header with non-200 OK status","file":"sr...
debug_error
def translate(configuration: Dict[str, Any], dictionary: Dict[str, str]) -> Dict[str, Any]: return { dictionary[field]: configuration[field] for field in dictionary if field in configuration }
def translate(configuration: Dict[str, Any], dictionary: Dict[str, str]) -> Dict[str, Any]: return {dictionary[field]: configuration[field] for field in dictionary}
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ray/anaconda3/bin/ray-operator", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ray/anaconda3/lib/python3.7/site-packages/ray/operator/operator.py", line 123, in main\\ncluster_config = operator_utils.cr_to_config(...
Traceback (most recent call last): File "/home/ray/anaconda3/bin/ray-operator", line 8, in <module> sys.exit(main()) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/operator/operator.py", line 123, in main cluster_config = operator_utils.cr_to_config(cluster_cr) File "/home/ray/anaconda3/lib/python3.7/site-pa...
KeyError
def __init__(self, conn_str: str = "", secure: bool = False, metadata: List[Tuple[str, str]] = None, connection_retries: int = 3): """Initializes the worker side grpc client. Args: conn_str: The host:port connection str...
def __init__(self, conn_str: str = "", secure: bool = False, metadata: List[Tuple[str, str]] = None, connection_retries: int = 3): """Initializes the worker side grpc client. Args: conn_str: The host:port connection str...
[{'piece_type': 'error message', 'piece_content': 'Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with:\\nstatus = StatusCode.UNAVAILABLE\\ndetails = "Received http2 header with status: 502"\\ndebug_error_string = "{"created":"@1610606646.739755640","description":"Receive...
Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "Received http2 header with status: 502" debug_error_string = "{"created":"@1610606646.739755640","description":"Received http2 :status header with non-200 OK status","file":"sr...
debug_error
def __init__(self, name: str, description: str = "", tag_keys: Optional[Tuple[str]] = None): if len(name) == 0: raise ValueError("Empty name is not allowed. " "Please provide a metric name.") self._name = nam...
def __init__(self, name: str, description: str = "", tag_keys: Optional[Tuple[str]] = None): if len(name) == 0: raise ValueError("Empty name is not allowed. " "Please provide a metric name.") self._name = nam...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ...
Traceback (most recent call last): File "test.py", line 8, in <module> count.record(1.0, {"b": 2}) File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record self._metric.record(value, tags=final_tags) File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record c_tags[tag_k.encode("as...
AttributeError
def set_default_tags(self, default_tags: Dict[str, str]): """Set default tags of metrics. Example: >>> # Note that set_default_tags returns the instance itself. >>> counter = Counter("name") >>> counter2 = counter.set_default_tags({"a": "b"}) >>> asse...
def set_default_tags(self, default_tags: Dict[str, str]): """Set default tags of metrics. Example: >>> # Note that set_default_tags returns the instance itself. >>> counter = Counter("name") >>> counter2 = counter.set_default_tags({"a": "b"}) >>> asse...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ...
Traceback (most recent call last): File "test.py", line 8, in <module> count.record(1.0, {"b": 2}) File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record self._metric.record(value, tags=final_tags) File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record c_tags[tag_k.encode("as...
AttributeError
def record(self, value: float, tags: dict = None) -> None: """Record the metric point of the metric. Args: value(float): The value to be recorded as a metric point. """ assert self._metric is not None if tags is not None: for val in tags.values(): ...
def record(self, value: float, tags: dict = None) -> None: """Record the metric point of the metric. Args: value(float): The value to be recorded as a metric point. """ assert self._metric is not None default_tag_copy = self._default_tags.copy() default_t...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nfrom ray.util.metrics import Count\\n\\nray.init()\\n\\ncount = Count("abc", tag_keys=("a", "b"))\\ncount.set_default_tags({"a": 1})\\ncount.record(1.0, {"b": 2})'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile ...
Traceback (most recent call last): File "test.py", line 8, in <module> count.record(1.0, {"b": 2}) File "/Users/eoakes/code/ray/python/ray/util/metrics.py", line 84, in record self._metric.record(value, tags=final_tags) File "python/ray/includes/metric.pxi", line 54, in ray._raylet.Metric.record c_tags[tag_k.encode("as...
AttributeError
def postprocess_advantages(policy, sample_batch, other_agent_batches=None, episode=None): # Stub serving backward compatibility. deprecation_warning( old="rllib.agents.a3c.a3c_tf_policy.postprocess_advantages", new...
def postprocess_advantages(policy, sample_batch, other_agent_batches=None, episode=None): completed = sample_batch[SampleBatch.DONES][-1] if completed: last_r = 0.0 else: next_state = [] for i in range(p...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def setup_mixins(policy, obs_space, action_space, config): ValueNetworkMixin.__init__(policy, obs_space, action_space, config) LearningRateSchedule.__init__(policy, config["lr"], config["lr_schedule"])
def setup_mixins(policy, obs_space, action_space, config): ValueNetworkMixin.__init__(policy) LearningRateSchedule.__init__(policy, config["lr"], config["lr_schedule"])
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def __init__(self, action_dist, actions, advantages, v_target, vf, vf_loss_coeff=0.5, entropy_coeff=0.01): log_prob = action_dist.logp(actions) # The "policy gradients" loss ...
def __init__(self): @make_tf_callable(self.get_session()) def value(ob, prev_action, prev_reward, *state): model_out, _ = self.model({ SampleBatch.CUR_OBS: tf.convert_to_tensor([ob]), SampleBatch.PREV_ACTIONS: tf.convert_to_tensor([prev_action]), ...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def add_advantages(policy, sample_batch, other_agent_batches=None, episode=None): # Stub serving backward compatibility. deprecation_warning( old="rllib.agents.a3c.a3c_torch_policy.add_advantages", new="rllib.evaluation.postprocessing.com...
def add_advantages(policy, sample_batch, other_agent_batches=None, episode=None): completed = sample_batch[SampleBatch.DONES][-1] if completed: last_r = 0.0 else: last_r = policy._value(sample_batch[SampleBatch.NEXT_OBS][-1]) ret...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def clip_gradients(policy: Policy, optimizer: "tf.keras.optimizers.Optimizer", loss: TensorType) -> ModelGradients: return minimize_and_clip( optimizer, loss, var_list=policy.q_func_vars, clip_val=policy.config["grad_clip"])
def clip_gradients(policy: Policy, optimizer: "tf.keras.optimizers.Optimizer", loss: TensorType) -> ModelGradients: if policy.config["grad_clip"] is not None: grads_and_vars = minimize_and_clip( optimizer, loss, var_list=policy.q_func_vars, ...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def get_policy_class(config): if config["framework"] == "torch": if config["vtrace"]: from ray.rllib.agents.impala.vtrace_torch_policy import \\ VTraceTorchPolicy return VTraceTorchPolicy else: from ray.rllib.agents.a3c.a3c_torch_policy import \\ ...
def get_policy_class(config): if config["framework"] == "torch": if config["vtrace"]: from ray.rllib.agents.impala.vtrace_torch_policy import \\ VTraceTorchPolicy return VTraceTorchPolicy else: from ray.rllib.agents.a3c.a3c_torch_policy import \\ ...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def postprocess_trajectory( policy: Policy, sample_batch: SampleBatch, other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None, episode: Optional[MultiAgentEpisode] = None) -> SampleBatch: """Postprocesses a trajectory and returns the processed trajectory. The trajector...
def postprocess_trajectory( policy: Policy, sample_batch: SampleBatch, other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None, episode: Optional[MultiAgentEpisode] = None) -> SampleBatch: """Postprocesses a trajectory and returns the processed trajectory. The trajector...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def postprocess_ppo_gae( policy: Policy, sample_batch: SampleBatch, other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None, episode: Optional[MultiAgentEpisode] = None) -> SampleBatch: # Stub serving backward compatibility. deprecation_warning( old="rllib.agent...
def postprocess_ppo_gae( policy: Policy, sample_batch: SampleBatch, other_agent_batches: Optional[Dict[AgentID, SampleBatch]] = None, episode: Optional[MultiAgentEpisode] = None) -> SampleBatch: """Postprocesses a trajectory and returns the processed trajectory. The trajectory c...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def gradients(self, optimizer, loss): self.gvs = { k: minimize_and_clip(optimizer, self.losses[k], self.vars[k], self.config["grad_norm_clipping"]) for k, optimizer in self.optimizers.items() } return self.gvs["critic"] + self.gvs["act...
def gradients(self, optimizer, loss): if self.config["grad_norm_clipping"] is not None: self.gvs = { k: minimize_and_clip(optimizer, self.losses[k], self.vars[k], self.config["grad_norm_clipping"]) for k, optimizer in self.opti...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def minimize_and_clip(optimizer, objective, var_list, clip_val=10.0): """Minimized `objective` using `optimizer` w.r.t. variables in `var_list` while ensure the norm of the gradients for each variable is clipped to `clip_val` """ # Accidentally passing values < 0.0 will break all gradients. asse...
def minimize_and_clip(optimizer, objective, var_list, clip_val=10.0): """Minimized `objective` using `optimizer` w.r.t. variables in `var_list` while ensure the norm of the gradients for each variable is clipped to `clip_val` """ # Accidentally passing values < 0.0 will break all gradients. asse...
[{'piece_type': 'error message', 'piece_content': '2020-06-21 13:09:32,571\\tERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event.\\nTraceback (most recent call last):\\nFile "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _pro...
2020-06-21 13:09:32,571 ERROR trial_runner.py:524 -- Trial PPO_TestingGym_f28cf_00000: Error processing event. Traceback (most recent call last): File "/home/user/anaconda3/envs/RLlibTesting/lib/python3.8/site-packages/ray/tune/trial_runner.py", line 472, in _process_trial result = self.trial_executor.fetch_result(tria...
tensorflow.python.framework.errors_impl.InvalidArgumentError
def persistent_id(self, obj): if isinstance(obj, ray.ObjectRef): obj_id = obj.binary() if obj_id not in self.server.object_refs[self.client_id]: # We're passing back a reference, probably inside a reference. # Let's hold onto it. self.s...
def persistent_id(self, obj): if isinstance(obj, ray.ObjectRef): obj_id = obj.binary() if obj_id not in self.server.object_refs[self.client_id]: # We're passing back a reference, probably inside a reference. # Let's hold onto it. self.s...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\n@ray.remote\\nclass B:\\npass\\n\\n@ray.remote\\nclass A:\\ndef __init__(self):\\nself.b = B.remote()\\n\\ndef get_actor_ref(self):\\nreturn [self.b]\\n\\na = A.remote()\\nray.get(a.get_actor_ref.remote())'}, {'piece_...
Got Error from data channel -- shutting down: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.UNKNOWN details = "Exception iterating responses: 'ServerPickler' object has no attribute 'actor_refs'" debug_error_string = "{"created":"@1610645797.119742000","description":"Error received from pee...
debug_error
def ClusterInfo(self, request, context=None) -> ray_client_pb2.ClusterInfoResponse: resp = ray_client_pb2.ClusterInfoResponse() resp.type = request.type if request.type == ray_client_pb2.ClusterInfoType.CLUSTER_RESOURCES: with disable_client_hook(): ...
def ClusterInfo(self, request, context=None) -> ray_client_pb2.ClusterInfoResponse: resp = ray_client_pb2.ClusterInfoResponse() resp.type = request.type if request.type == ray_client_pb2.ClusterInfoType.CLUSTER_RESOURCES: with disable_client_hook(): ...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\nprint(ray.get_runtime_context().node_id.hex())'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "test.py", line 4, in <module>\\nprint(ray.get_runtime_context().node_id.he...
Traceback (most recent call last): File "test.py", line 4, in <module> print(ray.get_runtime_context().node_id.hex()) File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 58, in node_id node_id = self.worker.current_node_id File "/Users/eoakes/code/ray/python/ray/worker.py", line 148, in current_node_id re...
AttributeError
def get_cluster_info(self, type: ray_client_pb2.ClusterInfoType.TypeEnum): req = ray_client_pb2.ClusterInfoRequest() req.type = type resp = self.server.ClusterInfo(req, metadata=self.metadata) if resp.WhichOneof("response_type") == "resource_table": # translate from a pro...
def get_cluster_info(self, type: ray_client_pb2.ClusterInfoType.TypeEnum): req = ray_client_pb2.ClusterInfoRequest() req.type = type resp = self.server.ClusterInfo(req, metadata=self.metadata) if resp.WhichOneof("response_type") == "resource_table": # translate from a pro...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nray.util.connect("localhost:61234")\\n\\nprint(ray.get_runtime_context().node_id.hex())'}, {'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "test.py", line 4, in <module>\\nprint(ray.get_runtime_context().node_id.he...
Traceback (most recent call last): File "test.py", line 4, in <module> print(ray.get_runtime_context().node_id.hex()) File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 58, in node_id node_id = self.worker.current_node_id File "/Users/eoakes/code/ray/python/ray/worker.py", line 148, in current_node_id re...
AttributeError
def get(self): """Get a dictionary of the current context. Fields that are not available (e.g., actor ID inside a task) won't be included in the field. Returns: dict: Dictionary of the current context. """ context = { "job_id": self.job_id, ...
def get(self): """Get a dictionary of the current_context. For fields that are not available (for example actor id inside a task) won't be included in the field. Returns: dict: Dictionary of the current context. """ context = { "job_id": self...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nray.init()\\n\\nprint(ray.get_runtime_context().get())'}, {'piece_type': 'error message', 'piece_content': '2021-01-13 14:15:29,011 INFO services.py:1169 -- View the Ray dashboard at http://127.0.0.1:8266\\nTraceback (most recent call last):\\nFile "test.py"...
2021-01-13 14:15:29,011 INFO services.py:1169 -- View the Ray dashboard at http://127.0.0.1:8266 Traceback (most recent call last): File "test.py", line 4, in <module> print(ray.get_runtime_context().get()) File "/Users/eoakes/code/ray/python/ray/runtime_context.py", line 26, in get "task_id": self.task_id, File "/User...
AssertionError
def __init__(self, node_ip_address, redis_address, dashboard_agent_port, redis_password=None, temp_dir=None, log_dir=None, metrics_export_port=None, node_manager_port=None, ...
def __init__(self, node_ip_address, redis_address, dashboard_agent_port, redis_password=None, temp_dir=None, log_dir=None, metrics_export_port=None, node_manager_port=None, ...
[{'piece_type': 'error message', 'piece_content': '(pid=None) Traceback (most recent call last):\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\new_dashboard/agent.py", line 317, in <module>\\n(pid=None) raise e\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\e...
(pid=None) Traceback (most recent call last): (pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 317, in <module> (pid=None) raise e (pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 293, in <modul...
KeyError
async def run(self): async def _check_parent(): """Check if raylet is dead and fate-share if it is.""" try: curr_proc = psutil.Process() while True: parent = curr_proc.parent() if (parent is None or parent.pid ==...
async def run(self): async def _check_parent(): """Check if raylet is dead and fate-share if it is.""" try: curr_proc = psutil.Process() while True: parent = curr_proc.parent() if (parent is None or parent.pid ==...
[{'piece_type': 'error message', 'piece_content': '(pid=None) Traceback (most recent call last):\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\new_dashboard/agent.py", line 317, in <module>\\n(pid=None) raise e\\n(pid=None) File "D:\\\\Programs\\\\Anaconda3\\\\e...
(pid=None) Traceback (most recent call last): (pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 317, in <module> (pid=None) raise e (pid=None) File "D:\\Programs\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\new_dashboard/agent.py", line 293, in <modul...
KeyError
def __str__(self): return ("The worker died unexpectedly while executing this task. " "Check python-core-worker-*.log files for more information.")
def __str__(self): return "The worker died unexpectedly while executing this task."
[{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n...
2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event. Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial result = self.trial_executor.fetch_result(trial) File "/usr...
ray.exceptions.RayTaskError
def __str__(self): return ("The actor died unexpectedly before finishing this task. " "Check python-core-worker-*.log files for more information.")
def __str__(self): return "The actor died unexpectedly before finishing this task."
[{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n...
2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event. Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial result = self.trial_executor.fetch_result(trial) File "/usr...
ray.exceptions.RayTaskError
def _try_to_compute_deterministic_class_id(cls, depth=5): """Attempt to produce a deterministic class ID for a given class. The goal here is for the class ID to be the same when this is run on different worker processes. Pickling, loading, and pickling again seems to produce more consistent results tha...
def _try_to_compute_deterministic_class_id(cls, depth=5): """Attempt to produce a deterministic class ID for a given class. The goal here is for the class ID to be the same when this is run on different worker processes. Pickling, loading, and pickling again seems to produce more consistent results tha...
[{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n...
2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event. Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial result = self.trial_executor.fetch_result(trial) File "/usr...
ray.exceptions.RayTaskError
def _random_string(): id_hash = hashlib.shake_128() id_hash.update(uuid.uuid4().bytes) id_bytes = id_hash.digest(ray_constants.ID_SIZE) assert len(id_bytes) == ray_constants.ID_SIZE return id_bytes
def _random_string(): id_hash = hashlib.sha1() id_hash.update(uuid.uuid4().bytes) id_bytes = id_hash.digest() assert len(id_bytes) == ray_constants.ID_SIZE return id_bytes
[{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n...
2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event. Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial result = self.trial_executor.fetch_result(trial) File "/usr...
ray.exceptions.RayTaskError
def run_function_on_all_workers(self, function, run_on_other_drivers=False): """Run arbitrary code on all of the workers. This function will first be run on the driver, and then it will be exported to all of the workers to be run. It will also be run on a...
def run_function_on_all_workers(self, function, run_on_other_drivers=False): """Run arbitrary code on all of the workers. This function will first be run on the driver, and then it will be exported to all of the workers to be run. It will also be run on a...
[{'piece_type': 'error message', 'piece_content': '2020-10-05 01:55:09,393\\\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event.\\nTraceback (most recent call last):\\nFile "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial\\n...
2020-10-05 01:55:09,393\\u0009ERROR trial_runner.py:567 -- Trial PPO_QbertNoFrameskip-v4_b43b9_00027: Error processing event. Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/trial_runner.py", line 515, in _process_trial result = self.trial_executor.fetch_result(trial) File "/usr...
ray.exceptions.RayTaskError
def _home(self): if self._home_cached is not None: return self._home_cached for _ in range(MAX_HOME_RETRIES - 1): try: self._home_cached = self._try_to_get_home() return self._home_cached except Exception: # TODO (Dm...
def _home(self): # TODO (Dmitri): Think about how to use the node's HOME variable # without making an extra kubectl exec call. if self._home_cached is None: cmd = self.kubectl + [ "exec", "-it", self.node_id, "--", "printenv", "HOME" ] join...
[{'piece_type': 'error message', 'piece_content': 'Updating cluster configuration. [hash=0194a452ebd82e0ab6eade7b4dd3a4f4f775d5de]\\nNew status: syncing-files\\n[2/7] Processing file mounts\\n2020-12-15 09:03:24,116\\tINFO command_runner.py:169 -- NodeUpdater: ray-head-m5nvd: Running kubectl -n ray exec -it ray-head-m5...
Updating cluster configuration. [hash=0194a452ebd82e0ab6eade7b4dd3a4f4f775d5de] New status: syncing-files [2/7] Processing file mounts 2020-12-15 09:03:24,116 INFO command_runner.py:169 -- NodeUpdater: ray-head-m5nvd: Running kubectl -n ray exec -it ray-head-m5nvd -- bash --login -c -i 'true &amp;&amp; source ~/.bashrc...
subprocess.CalledProcessError
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "debugging.py", line 2, in <module>\\nray.init()\\nFile "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init\\nray_params=ray_params)\\nFile "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init...
Traceback (most recent call last): File "debugging.py", line 2, in <module> ray.init() File "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init ray_params=ray_params) File "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init__ self.start_head_processes() File "/Users/haochen/code/ant_...
redis.exceptions.ResponseError
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "debugging.py", line 2, in <module>\\nray.init()\\nFile "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init\\nray_params=ray_params)\\nFile "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init...
Traceback (most recent call last): File "debugging.py", line 2, in <module> ray.init() File "/Users/haochen/code/ant_ray/python/ray/worker.py", line 740, in init ray_params=ray_params) File "/Users/haochen/code/ant_ray/python/ray/node.py", line 200, in __init__ self.start_head_processes() File "/Users/haochen/code/ant_...
redis.exceptions.ResponseError
def shutdown(self) -> None: """Completely shut down the connected Serve instance. Shuts down all processes and deletes all state associated with the instance. """ if (not self._shutdown) and ray.is_initialized(): ray.get(self._controller.shutdown.remote()) ...
def shutdown(self) -> None: """Completely shut down the connected Serve instance. Shuts down all processes and deletes all state associated with the instance. """ if not self._shutdown: ray.get(self._controller.shutdown.remote()) ray.kill(self._contro...
[{'piece_type': 'source code', 'piece_content': 'from ray import serve\\n\\nclient = serve.start(detached=True)\\nclient.shutdown()\\nclient = serve.start(detached=True)'}, {'piece_type': 'error message', 'piece_content': 'File descriptor limit 256 is too low for production servers and may result in connection errors. ...
File descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with 'ulimit -n 8192' 2020-11-20 10:38:27,065 INFO services.py:1173 -- View the Ray dashboard at http://127.0.0.1:8265 (pid=raylet) 2020-11-20 10:38:29,628 INFO controller.py:313 -- Sta...
ray.serve.exceptions.RayServeException
def __init__( self, obs_space: gym.spaces.Space, action_space: gym.spaces.Space, config: TrainerConfigDict, loss_fn: Callable[[ Policy, ModelV2, Type[TFActionDistribution], SampleBatch ], TensorType], *, ...
def __init__( self, obs_space: gym.spaces.Space, action_space: gym.spaces.Space, config: TrainerConfigDict, loss_fn: Callable[[ Policy, ModelV2, Type[TFActionDistribution], SampleBatch ], TensorType], *, ...
[{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi...
WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version. Instructions for updating: This function will only be available ...
tensorflow.python.ops.op_selector.UnliftableError
def __init__(self, observation_space: gym.spaces.Space, action_space: gym.spaces.Space, config: TrainerConfigDict, sess: "tf1.Session", obs_input: TensorType, sampled_action: TensorType, loss: Tens...
def __init__(self, observation_space: gym.spaces.Space, action_space: gym.spaces.Space, config: TrainerConfigDict, sess: "tf1.Session", obs_input: TensorType, sampled_action: TensorType, loss: Tens...
[{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi...
WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version. Instructions for updating: This function will only be available ...
tensorflow.python.ops.op_selector.UnliftableError
def try_import_tf(error=False): """Tries importing tf and returns the module (or None). Args: error (bool): Whether to raise an error if tf cannot be imported. Returns: Tuple: - tf1.x module (either from tf2.x.compat.v1 OR as tf1.x). - tf module (resulting from `imp...
def try_import_tf(error=False): """Tries importing tf and returns the module (or None). Args: error (bool): Whether to raise an error if tf cannot be imported. Returns: Tuple: - tf1.x module (either from tf2.x.compat.v1 OR as tf1.x). - tf module (resulting from `imp...
[{'piece_type': 'other', 'piece_content': 'from gym.spaces import Discrete\\n\\nimport ray\\nfrom ray.rllib.examples.env.rock_paper_scissors import RockPaperScissors\\nfrom ray.rllib.agents import ppo\\n\\n\\nselect_policy = lambda agent_id: "policy_01" if agent_id == "player1" else "policy_02"\\n\\nconfig = {\\n"multi...
WARNING:tensorflow:From /Users/ivallesp/projects/rockpaperscisors/.venv/lib/python3.7/site-packages/ray/rllib/policy/tf_policy.py:653: build_tensor_info (from tensorflow.python.saved_model.utils_impl) is deprecated and will be removed in a future version. Instructions for updating: This function will only be available ...
tensorflow.python.ops.op_selector.UnliftableError
def build_trainer( name: str, *, default_config: Optional[TrainerConfigDict] = None, validate_config: Optional[Callable[[TrainerConfigDict], None]] = None, default_policy: Optional[Type[Policy]] = None, get_policy_class: Optional[Callable[[TrainerConfigDict], Optional[Typ...
def build_trainer( name: str, *, default_config: Optional[TrainerConfigDict] = None, validate_config: Optional[Callable[[TrainerConfigDict], None]] = None, default_policy: Optional[Type[Policy]] = None, get_policy_class: Optional[Callable[[TrainerConfigDict], Optional[Typ...
[{'piece_type': 'other', 'piece_content': 'from or_gym.utils import create_env\\nfrom gym import spaces\\nfrom ray.rllib.utils import try_import_tf\\nfrom ray.rllib.models.tf.fcnet import FullyConnectedNetwork\\nfrom ray.rllib.models.tf.tf_modelv2 import TFModelV2\\nfrom ray.rllib.models import ModelCatalog\\nfrom ray ...
Traceback (most recent call last): File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/ray/rllib/models/preprocessors.py", line 60, in check_shape if not self._obs_space.contains(observation): File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/gym/spaces/box.py", line 128, in co...
AttributeError
def __init__(self, *, env_creator: Optional[Callable[[EnvContext], EnvType]] = None, validate_env: Optional[Callable[[EnvType], None]] = None, policy_class: Optional[Type[Policy]] = None, trainer_config: Optional[TrainerConfigDict]...
def __init__(self, *, env_creator: Optional[Callable[[EnvContext], EnvType]] = None, validate_env: Optional[Callable[[EnvType], None]] = None, policy_class: Optional[Type[Policy]] = None, trainer_config: Optional[TrainerConfigDict]...
[{'piece_type': 'other', 'piece_content': 'from or_gym.utils import create_env\\nfrom gym import spaces\\nfrom ray.rllib.utils import try_import_tf\\nfrom ray.rllib.models.tf.fcnet import FullyConnectedNetwork\\nfrom ray.rllib.models.tf.tf_modelv2 import TFModelV2\\nfrom ray.rllib.models import ModelCatalog\\nfrom ray ...
Traceback (most recent call last): File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/ray/rllib/models/preprocessors.py", line 60, in check_shape if not self._obs_space.contains(observation): File "/data2/huangcq/miniconda3/envs/majenv/lib/python3.8/site-packages/gym/spaces/box.py", line 128, in co...
AttributeError
def memory_summary(): """Returns a formatted string describing memory usage in the cluster.""" import grpc from ray.core.generated import node_manager_pb2 from ray.core.generated import node_manager_pb2_grpc # We can ask any Raylet for the global memory info. raylet = ray.nodes()[0] raylet...
def memory_summary(): """Returns a formatted string describing memory usage in the cluster.""" import grpc from ray.core.generated import node_manager_pb2 from ray.core.generated import node_manager_pb2_grpc # We can ask any Raylet for the global memory info. raylet = ray.nodes()[0] raylet...
[{'piece_type': 'error message', 'piece_content': '2020-05-19 02:13:32,283\\tINFO scripts.py:976 -- Connecting to Ray instance at 172.31.6.12:34940.\\n2020-05-19 02:13:32,284\\tWARNING worker.py:809 -- When connecting to an existing cluster, _internal_config must match the cluster\\'s _internal_config.\\n(pid=5906) E05...
2020-05-19 02:13:32,283 INFO scripts.py:976 -- Connecting to Ray instance at 172.31.6.12:34940. 2020-05-19 02:13:32,284 WARNING worker.py:809 -- When connecting to an existing cluster, _internal_config must match the cluster's _internal_config. (pid=5906) E0519 02:13:32.383447 5906 plasma_store_provider.cc:108] Failed...
grpc._channel._InactiveRpcError
def custom_excepthook(type, value, tb): # If this is a driver, push the exception to GCS worker table. if global_worker.mode == SCRIPT_MODE: error_message = "".join(traceback.format_tb(tb)) worker_id = global_worker.worker_id worker_type = ray.gcs_utils.DRIVER worker_info = {"exc...
def custom_excepthook(type, value, tb): # If this is a driver, push the exception to GCS worker table. if global_worker.mode == SCRIPT_MODE: error_message = "".join(traceback.format_tb(tb)) worker_id = global_worker.worker_id worker_type = ray.gcs_utils.DRIVER worker_info = {"exc...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nray.init()\\n\\nraise Exception("foobar")'}, {'piece_type': 'error message', 'piece_content': 'raise Exception("hello")\\nError in sys.excepthook:\\nTraceback (most recent call last):\\nFile "/Users/eoakes/code/ray/python/ray/worker.py", line 836, in custom_...
raise Exception("hello") Error in sys.excepthook: Traceback (most recent call last): File "/Users/eoakes/code/ray/python/ray/worker.py", line 836, in custom_excepthook ray.state.state.add_worker(worker_id, worker_type, worker_info) File "/Users/eoakes/code/ray/python/ray/state.py", line 733, in add_worker return self.g...
AttributeError
def __init__(self, sync_up_template, sync_down_template, delete_template=noop_template): """Syncs between two directories with the given command. Arguments: sync_up_template (str): A runnable string template; needs to in...
def __init__(self, sync_up_template, sync_down_template, delete_template=noop_template): """Syncs between two directories with the given command. Arguments: sync_up_template (str): A runnable string template; needs to in...
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
def set_logdir(self, logdir): """Sets the directory to log sync execution output in. Args: logdir (str): Log directory. """ self.logfile = tempfile.NamedTemporaryFile( prefix="log_sync_out", dir=logdir, suffix=".log", delete=False) self._closed = Fals...
def set_logdir(self, logdir): """Sets the directory to log sync execution output in. Args: logdir (str): Log directory. """ self.logfile = tempfile.NamedTemporaryFile( prefix="log_sync_out", dir=logdir, suffix=".log", delete=False)
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
def delete(self, target): if self.is_running: logger.warning("Last sync client cmd still in progress, skipping.") return False final_cmd = self.delete_template.format(target=quote(target)) logger.debug("Running delete: {}".format(final_cmd)) self.cmd_process =...
def delete(self, target): if self.is_running: logger.warning("Last sync client cmd still in progress, skipping.") return False final_cmd = self.delete_template.format(target=quote(target)) logger.debug("Running delete: {}".format(final_cmd)) self.cmd_process =...
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
def _execute(self, sync_template, source, target): """Executes sync_template on source and target.""" if self.is_running: logger.warning("Last sync client cmd still in progress, skipping.") return False final_cmd = sync_template.format( source=quote(source...
def _execute(self, sync_template, source, target): """Executes sync_template on source and target.""" if self.is_running: logger.warning("Last sync client cmd still in progress, skipping.") return False final_cmd = sync_template.format( source=quote(source...
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
def on_trial_complete(self, iteration: int, trials: List["Trial"], trial: "Trial", **info): trial_syncer = self._get_trial_syncer(trial) if NODE_IP in trial.last_result: trainable_ip = trial.last_result[NODE_IP] else: trainable_ip = ray.get(t...
def on_trial_complete(self, iteration: int, trials: List["Trial"], trial: "Trial", **info): trial_syncer = self._get_trial_syncer(trial) if NODE_IP in trial.last_result: trainable_ip = trial.last_result[NODE_IP] else: trainable_ip = ray.get(t...
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
def mock_storage_client(): """Mocks storage client that treats a local dir as durable storage.""" client = get_sync_client(LOCAL_SYNC_TEMPLATE, LOCAL_DELETE_TEMPLATE) path = os.path.join(ray.utils.get_user_temp_dir(), f"mock-client-{uuid.uuid4().hex[:4]}") os.makedirs(path, exist...
def mock_storage_client(): """Mocks storage client that treats a local dir as durable storage.""" return get_sync_client(LOCAL_SYNC_TEMPLATE, LOCAL_DELETE_TEMPLATE)
[{'piece_type': 'error message', 'piece_content': '2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down\\nself._local_dir)\\nFile "/home/ubunt...
2020-11-21 02:02:05,077 ERROR syncer.py:190 -- Sync execution failed. Traceback (most recent call last): File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-packages/ray/tune/syncer.py", line 187, in sync_down self._local_dir) File "/home/ubuntu/anaconda3/envs/tensorflow2_p36/lib/python3.6/site-package...
OSError
async def _do_long_poll(self): while True: try: updates: Dict[str, UpdatedObject] = await self._poll_once() self._update(updates) logger.debug(f"LongPollerClient received udpates: {updates}") for key, updated_object in updates.items...
async def _do_long_poll(self): while True: updates: Dict[str, UpdatedObject] = await self._poll_once() self._update(updates) logger.debug(f"LongPollerClient received updates: {updates}") for key, updated_object in updates.items(): # NOTE(simon)...
[{'piece_type': 'source code', 'piece_content': 'import ray\\nimport asyncio\\n\\nray.init()\\n@ray.remote\\nclass A:\\ndef hi(self): return "hi"\\n\\na = A.remote()\\nasync def main():\\nawait a.hi.remote()\\nray.kill(a)\\nawait a.hi.remote()\\n\\nasyncio.get_event_loop().run_until_complete(main())'}, {'piece_type': '...
Exception in callback async_set_result.<locals>.set_future() handle: <Handle async_set_result.<locals>.set_future()> Traceback (most recent call last): File "/Users/simonmo/miniconda3/lib/python3.6/asyncio/events.py", line 145, in _run self._callback(*self._args) File "python/ray/_raylet.pyx", line 1530, in ray._raylet...
AttributeError
async def _get_actor(actor): actor = dict(actor) worker_id = actor["address"]["workerId"] core_worker_stats = DataSource.core_worker_stats.get(worker_id, {}) actor_constructor = core_worker_stats.get("actorTitle", "Unknown actor const...
async def _get_actor(actor): actor = dict(actor) worker_id = actor["address"]["workerId"] core_worker_stats = DataSource.core_worker_stats.get(worker_id, {}) actor_constructor = core_worker_stats.get("actorTitle", "Unknown actor const...
[{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 347, in _update_cache response = task.result() File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_...
Error: Traceback (most recent call last): File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 347, in _update_cache response = task.result() File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 77, in get...
TypeError
def get_address_info_from_redis_helper(redis_address, node_ip_address, redis_password=None): redis_ip_address, redis_port = redis_address.split(":") # Get node table from global state accessor. global_state = ray.state.GlobalState...
def get_address_info_from_redis_helper(redis_address, node_ip_address, redis_password=None): redis_ip_address, redis_port = redis_address.split(":") # Get node table from global state accessor. global_state = ray.state.GlobalState...
[{'piece_type': 'error message', 'piece_content': '020-11-11 14:48:28,960\\tINFO worker.py:672 -- Connecting to existing Ray cluster at address: ***:***\\n2020-11-11 14:48:28,968\\tWARNING services.py:218 -- Some processes that the driver needs to connect to have not registered with Redis, so retrying. Have you run \\'...
020-11-11 14:48:28,960 INFO worker.py:672 -- Connecting to existing Ray cluster at address: ***:*** 2020-11-11 14:48:28,968 WARNING services.py:218 -- Some processes that the driver needs to connect to have not registered with Redis, so retrying. Have you run 'ray start' on this node? 2020-11-11 14:48:29,977 WARNING se...
RuntimeError
def __init__(self): self.indent_level = 0 self._verbosity = 0 self._verbosity_overriden = False self._color_mode = "auto" self._log_style = "record" self.pretty = False self.interactive = False # store whatever colorful has detected for future use if...
def __init__(self): self.indent_level = 0 self._verbosity = 0 self._color_mode = "auto" self._log_style = "record" self.pretty = False self.interactive = False # store whatever colorful has detected for future use if # the color ouput is toggled (col...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def verbosity(self): if self._verbosity_overriden: return self._verbosity elif not self.pretty: return 999 return self._verbosity
def verbosity(self): if not self.pretty: return 999 return self._verbosity
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def _set_verbosity(self, x): self._verbosity = x self._verbosity_overriden = True
def _set_verbosity(self, x): self._verbosity = x
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def rsync(config_file: str, source: Optional[str], target: Optional[str], override_cluster_name: Optional[str], down: bool, ip_address: Optional[str] = None, use_internal_ip: bool = False, no_config_cache: bool = False, all_nodes: bool = Fa...
def rsync(config_file: str, source: Optional[str], target: Optional[str], override_cluster_name: Optional[str], down: bool, ip_address: Optional[str] = None, use_internal_ip: bool = False, no_config_cache: bool = False, all_nodes: bool = Fa...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def rsync_to_node(node_id, is_head_node): updater = NodeUpdaterThread( node_id=node_id, provider_config=config["provider"], provider=provider, auth_config=config["auth"], cluster_name=config["cluster_name"], file_mounts=config["file_mou...
def rsync_to_node(node_id, is_head_node): updater = NodeUpdaterThread( node_id=node_id, provider_config=config["provider"], provider=provider, auth_config=config["auth"], cluster_name=config["cluster_name"], file_mounts=config["file_mou...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def __init__(self, local_dir: str, remote_dir: str, sync_client: Optional[SyncClient] = None): configure_logging( log_style="record", verbosity=env_integer("TUNE_SYNCER_VERBOSITY", 0)) self.local_ip = services.get_node_ip_add...
def __init__(self, local_dir: str, remote_dir: str, sync_client: Optional[SyncClient] = None): self.local_ip = services.get_node_ip_address() self.worker_ip = None sync_client = sync_client or DockerSyncClient() sync_client.configur...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def reset_trial(self, trial, new_config, new_experiment_tag, logger_creator=None): """Tries to invoke `Trainable.reset()` to reset trial. Args: trial (Trial): Trial to be reset. new_config (dict)...
def reset_trial(self, trial, new_config, new_experiment_tag, logger_creator=None): """Tries to invoke `Trainable.reset()` to reset trial. Args: trial (Trial): Trial to be reset. new_config (dict)...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def reset(self, new_config, logger_creator=None): """Resets trial for use with new config. Subclasses should override reset_config() to actually reset actor behavior for the new config.""" self.config = new_config self._result_logger.flush() self._result_logger.clos...
def reset(self, new_config, logger_creator=None): """Resets trial for use with new config. Subclasses should override reset_config() to actually reset actor behavior for the new config.""" self.config = new_config self._result_logger.flush() self._result_logger.clos...
[{'piece_type': 'error message', 'piece_content': '2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn\\nist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64...
2020-11-19 19:42:27,243 VINFO updater.py:460 -- `rsync`ed /root/ray_results/tune_mnist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (remote) to /root/ray_results/tune_mn ist_pbt/DEFAULT_1d092_00006_6_layer_1_size=128,layer_2_size=64_2020-11-19_19-41-58/ (local) 2020-11-19 19:42:27,247...
ray.tune.error.TuneError
def _bootstrap_config(config: Dict[str, Any], no_config_cache: bool = False) -> Dict[str, Any]: config = prepare_config(config) hasher = hashlib.sha1() hasher.update(json.dumps([config], sort_keys=True).encode("utf-8")) cache_key = os.path.join(tempfile.gettempdir(), ...
def _bootstrap_config(config: Dict[str, Any], no_config_cache: bool = False) -> Dict[str, Any]: config = prepare_config(config) hasher = hashlib.sha1() hasher.update(json.dumps([config], sort_keys=True).encode("utf-8")) cache_key = os.path.join(tempfile.gettempdir(), ...
[{'piece_type': 'other', 'piece_content': '(base) ➜ tune git:(fix-kubernetes-dep) ✗ pip list | grep boto\\nboto 2.49.0\\nboto3 1.4.8\\nbotocore 1.8.50'}, {'piece_type': 'error message', 'piece_content': '(base) ➜ tune git:(fix-kubernetes-dep) ✗ ray...
(base) ➜ tune git:(fix-kubernetes-dep) ✗ ray up $CFG -y Cluster: basic Checking AWS environment settings Traceback (most recent call last): File "/Users/rliaw/miniconda3/bin/ray", line 8, in <module> sys.exit(main()) File "/Users/rliaw/miniconda3/lib/python3.7/site-packages/ray/scripts/scripts.py", line 1471, in main...
AttributeError
async def get_node_workers(cls, node_id): workers = [] node_ip = DataSource.node_id_to_ip[node_id] node_logs = DataSource.ip_and_pid_to_logs.get(node_ip, {}) node_errs = DataSource.ip_and_pid_to_errors.get(node_ip, {}) node_physical_stats = DataSource.node_physical_stats.get(...
async def get_node_workers(cls, node_id): workers = [] node_ip = DataSource.node_id_to_ip[node_id] node_logs = DataSource.ip_and_pid_to_logs.get(node_ip, {}) logger.error(node_logs) node_errs = DataSource.ip_and_pid_to_errors.get(node_ip, {}) logger.error(node_errs) ...
[{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()\\nFile "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/st...
Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result() File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 89, in get_all_nodes all...
KeyError
async def get_node_info(cls, node_id): node_physical_stats = dict( DataSource.node_physical_stats.get(node_id, {})) node_stats = dict(DataSource.node_stats.get(node_id, {})) node = DataSource.nodes.get(node_id, {}) node_ip = DataSource.node_id_to_ip.get(node_id) #...
async def get_node_info(cls, node_id): node_physical_stats = dict( DataSource.node_physical_stats.get(node_id, {})) node_stats = dict(DataSource.node_stats.get(node_id, {})) node = DataSource.nodes.get(node_id, {}) node_ip = DataSource.node_id_to_ip.get(node_id) #...
[{'piece_type': 'error message', 'piece_content': 'Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result()\\nFile "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/st...
Error: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/utils.py", line 351, in _update_cache response = task.result() File "/root/anaconda3/lib/python3.7/site-packages/ray/new_dashboard/modules/stats_collector/stats_collector_head.py", line 89, in get_all_nodes all...
KeyError
def unflatten_dict(dt, delimiter="/"): """Unflatten dict. Does not support unflattening lists.""" dict_type = type(dt) out = dict_type() for key, val in dt.items(): path = key.split(delimiter) item = out for k in path[:-1]: item = item.setdefault(k, dict_type()) ...
def unflatten_dict(dt, delimiter="/"): """Unflatten dict. Does not support unflattening lists.""" out = defaultdict(dict) for key, val in dt.items(): path = key.split(delimiter) item = out for k in path[:-1]: item = item[k] item[path[-1]] = val return dict(out...
[{'piece_type': 'error message', 'piece_content': '$ python ./python/ray/tune/examples/bohb_example.py\\nFile descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with \\'ulimit -n 8192\\'\\n2020-11-11 20:38:44,944 INFO services.py:1110 -- Vie...
$ python ./python/ray/tune/examples/bohb_example.py File descriptor limit 256 is too low for production servers and may result in connection errors. At least 8192 is recommended. --- Fix with 'ulimit -n 8192' 2020-11-11 20:38:44,944 INFO services.py:1110 -- View the Ray dashboard at http://127.0.0.1:8265 Traceback (mos...
KeyError
def find_redis_address(address=None): pids = psutil.pids() redis_addresses = set() for pid in pids: try: proc = psutil.Process(pid) # HACK: Workaround for UNIX idiosyncrasy # Normally, cmdline() is supposed to return the argument list. # But it in some...
def find_redis_address(address=None): pids = psutil.pids() redis_addresses = set() for pid in pids: try: proc = psutil.Process(pid) # HACK: Workaround for UNIX idiosyncrasy # Normally, cmdline() is supposed to return the argument list. # But it in some...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/swang/anaconda3/envs/ray-36/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1462, in main\\nreturn cli()\\nFile "/home/...
Traceback (most recent call last): File "/home/swang/anaconda3/envs/ray-36/bin/ray", line 8, in <module> sys.exit(main()) File "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1462, in main return cli() File "/home/swang/anaconda3/envs/ray-36/lib/python3.6/site-packages/click...
ConnectionError
def _run_helper(self, final_cmd, with_output=False, exit_on_fail=False, silent=False): """Run a command that was already setup with SSH and `bash` settings. Args: cmd (List[str]): Full comman...
def _run_helper(self, final_cmd, with_output=False, exit_on_fail=False, silent=False): """Run a command that was already setup with SSH and `bash` settings. Args: cmd (List[str]): Full comman...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "XXX/lib/python3.7/site-packages/ray/autoscaler/command_runner.py", line 248, in run\\nself.process_runner.check_call(final_cmd, shell=True)\\nFile "/Users/mkoh/.pyenv/versions/3.7.7/lib/python3.7/subprocess.py", line 363, in c...
Traceback (most recent call last): File "XXX/lib/python3.7/site-packages/ray/autoscaler/command_runner.py", line 248, in run self.process_runner.check_call(final_cmd, shell=True) File "/Users/mkoh/.pyenv/versions/3.7.7/lib/python3.7/subprocess.py", line 363, in check_call raise CalledProcessError(retcode, cmd) subproce...
subprocess.CalledProcessError
def __init__(self, space: Optional[Union[Dict, List[Dict]]] = None, metric: Optional[str] = None, mode: Optional[str] = None, parameter_constraints: Optional[List] = None, outcome_constraints: Optional[List] = None, ...
def __init__(self, space: Optional[List[Dict]] = None, metric: Optional[str] = None, mode: Optional[str] = None, parameter_constraints: Optional[List] = None, outcome_constraints: Optional[List] = None, ax_client: ...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, space: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, utility_kwargs: Optional[Dict] = None, random_state: int = 42, random_search_steps: int = 10, ...
def __init__(self, space: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, utility_kwargs: Optional[Dict] = None, random_state: int = 42, random_search_steps: int = 10, ...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def convert_search_space(spec: Dict, join: bool = False) -> Dict: spec = flatten_dict(spec, prevent_delimiter=True) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if grid_vars: raise ValueError( "Grid search parameters cannot be automatically conve...
def convert_search_space(spec: Dict) -> Dict: spec = flatten_dict(spec, prevent_delimiter=True) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if grid_vars: raise ValueError( "Grid search parameters cannot be automatically converted " ...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, space: Optional[Union[Dict, ConfigSpace.ConfigurationSpace]] = None, bohb_config: Optional[Dict] = None, max_concurrent: int = 10, metric: Optional[str] = None, mode: Option...
def __init__(self, space: Optional[ConfigSpace.ConfigurationSpace] = None, bohb_config: Optional[Dict] = None, max_concurrent: int = 10, metric: Optional[str] = None, mode: Optional[str] = None): from hpbandster.optimizers....
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, optimizer: Optional[BlackboxOptimiser] = None, domain: Optional[str] = None, space: Optional[Union[Dict, List[Dict]]] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_e...
def __init__(self, optimizer: Optional[BlackboxOptimiser] = None, domain: Optional[str] = None, space: Optional[List[Dict]] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_evaluate: Opti...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__( self, space: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_evaluate: Optional[List[Dict]] = None, n_initial_points: int = 20, random_state_seed: Optional[int] = None, ...
def __init__( self, space: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_evaluate: Optional[List[Dict]] = None, n_initial_points: int = 20, random_state_seed: Optional[int] = None, ...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, optimizer: Union[None, Optimizer, ConfiguredOptimizer] = None, space: Optional[Union[Dict, Parameter]] = None, metric: Optional[str] = None, mode: Optional[str] = None, max_concurrent: Optional[int] = None, ...
def __init__(self, optimizer: Union[None, Optimizer, ConfiguredOptimizer] = None, space: Optional[Parameter] = None, metric: Optional[str] = None, mode: Optional[str] = None, max_concurrent: Optional[int] = None, *...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, space: Optional[Union[Dict, List[Tuple]]] = None, metric: Optional[str] = None, mode: Optional[str] = None, sampler: Optional[BaseSampler] = None): assert ot is not None, ( "Optuna must be installed! Run `pip ...
def __init__(self, space: Optional[List[Tuple]] = None, metric: Optional[str] = None, mode: Optional[str] = None, sampler: Optional[BaseSampler] = None): assert ot is not None, ( "Optuna must be installed! Run `pip install optun...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, optimizer: Optional[sko.optimizer.Optimizer] = None, space: Union[List[str], Dict[str, Union[Tuple, List]]] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_evaluate: Optional[List[List...
def __init__(self, optimizer: Optional[sko.optimizer.Optimizer] = None, space: Union[List[str], Dict[str, Union[Tuple, List]]] = None, metric: Optional[str] = None, mode: Optional[str] = None, points_to_evaluate: Optional[List[List...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def convert_search_space(spec: Dict, join: bool = False) -> Dict: spec = flatten_dict(spec, prevent_delimiter=True) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if grid_vars: raise ValueError( "Grid search parameters cannot be automatically conve...
def convert_search_space(spec: Dict) -> Dict: spec = flatten_dict(spec, prevent_delimiter=True) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if grid_vars: raise ValueError( "Grid search parameters cannot be automatically converted " ...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def __init__(self, algo: str = "asracos", budget: Optional[int] = None, dim_dict: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, **kwargs): assert zoopt is not None, "ZOOpt...
def __init__(self, algo: str = "asracos", budget: Optional[int] = None, dim_dict: Optional[Dict] = None, metric: Optional[str] = None, mode: Optional[str] = None, **kwargs): assert zoopt is not None, "ZOOpt...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def convert_search_space(spec: Dict, join: bool = False) -> Dict[str, Tuple]: spec = copy.deepcopy(spec) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if not domain_vars and not grid_vars: return {} if grid_vars: ...
def convert_search_space(spec: Dict) -> Dict[str, Tuple]: spec = copy.deepcopy(spec) resolved_vars, domain_vars, grid_vars = parse_spec_vars(spec) if not domain_vars and not grid_vars: return [] if grid_vars: raise ValueError( "Grid search pa...
[{'piece_type': 'other', 'piece_content': 'config = {\\n"learning_rate": tune.qloguniform(1e-4, 1e-1, 5e-5),\\n"batch_size": tune.choice([32, 64, 128, 256]),\\n"neurons1": tune.qrandint(32, 1024, 32),\\n"neurons2": tune.qrandint(32, 1024, 32),\\n"dropout": tune.choice([0.1, 0.2, 0.3,]),\\n}'}, {'piece_type': 'other', '...
Traceback (most recent call last): (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/threading.py", line 916, in _bootstrap_inner (pid=82997) self.run() (pid=82997) File "/home/gsukhorukov/.conda/envs/tf2/lib/python3.6/site-packages/ray/tune/function_runner.py", line 246, in run (pid=82997) ...
TypeError
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
def init( address=None, *, num_cpus=None, num_gpus=None, resources=None, object_store_memory=None, local_mode=False, ignore_reinit_error=False, include_dashboard=None, dashboard_host=ray_constants.DEFAULT_DASHBOARD_IP, dashboard_por...
[{'piece_type': 'other', 'piece_content': "(pid=136534) F1009 07:02:26.344231 136534 136534 service_based_gcs_client.cc:207] Couldn't reconnect to GCS server. The last attempted GCS server address was 192.168.1.98:39953\\n(pid=136534) *** Check failure stack trace: ***\\n(pid=136534) @ 0x7f56afdb7f5d google::L...
^CTraceback (most recent call last): File "test.py", line 72, in <module> **tune_kwargs) File "/home/demattia/miniconda3/envs/test_tune/lib/python3.7/site-packages/ray/tune/tune.py", line 405, in run runner.step() File "/home/demattia/miniconda3/envs/test_tune/lib/python3.7/site-packages/ray/tune/trial_runner.py", line...
CError
def with_parameters(fn, **kwargs): """Wrapper for function trainables to pass arbitrary large data objects. This wrapper function will store all passed parameters in the Ray object store and retrieve them when calling the function. It can thus be used to pass arbitrary data, even datasets, to Tune trai...
def with_parameters(fn, **kwargs): """Wrapper for function trainables to pass arbitrary large data objects. This wrapper function will store all passed parameters in the Ray object store and retrieve them when calling the function. It can thus be used to pass arbitrary data, even datasets, to Tune trai...
[{'piece_type': 'error message', 'piece_content': 'Failure # 1 (occurred at 2020-09-26_16-50-01)\\nTraceback (most recent call last):\\nFile "/home/karol/PycharmProjects/ray/python/ray/tune/trial_runner.py", line 518, in _process_trial\\nresult = self.trial_executor.fetch_result(trial)\\nFile "/home/karol/PycharmProjec...
Failure # 1 (occurred at 2020-09-26_16-50-01) Traceback (most recent call last): File "/home/karol/PycharmProjects/ray/python/ray/tune/trial_runner.py", line 518, in _process_trial result = self.trial_executor.fetch_result(trial) File "/home/karol/PycharmProjects/ray/python/ray/tune/ray_trial_executor.py", line 488, in...
ray.tune.error.TuneError
def wandb_mixin(func: Callable): """wandb_mixin Weights and biases (https://www.wandb.com/) is a tool for experiment tracking, model optimization, and dataset versioning. This Ray Tune Trainable mixin helps initializing the Wandb API for use with the ``Trainable`` class or with `@wandb_mixin` for t...
def wandb_mixin(func: Callable): """wandb_mixin Weights and biases (https://www.wandb.com/) is a tool for experiment tracking, model optimization, and dataset versioning. This Ray Tune Trainable mixin helps initializing the Wandb API for use with the ``Trainable`` class or with `@wandb_mixin` for t...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ubuntu/run_ray_tune.py", line 222, in <module>\\ntune_helsinki_(args)\\nFile "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_\\nray_wandb_func = wandb_mixin(ray_func)\\nFile "/home/ubuntu/anaconda3/lib/python3...
Traceback (most recent call last): File "/home/ubuntu/run_ray_tune.py", line 222, in <module> tune_helsinki_(args) File "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_ ray_wandb_func = wandb_mixin(ray_func) File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/tune/integration/wandb.py", line 142, in...
AttributeError
def _init(self): config = self.config.copy() config.pop("callbacks", None) # Remove callbacks try: if config.get("logger_config", {}).get("wandb"): logger_config = config.pop("logger_config") wandb_config = logger_config.get("wandb").copy() ...
def _init(self): config = self.config.copy() config.pop("callbacks", None) # Remove callbacks try: if config.get("logger_config", {}).get("wandb"): logger_config = config.pop("logger_config") wandb_config = logger_config.get("wandb").copy() ...
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "/home/ubuntu/run_ray_tune.py", line 222, in <module>\\ntune_helsinki_(args)\\nFile "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_\\nray_wandb_func = wandb_mixin(ray_func)\\nFile "/home/ubuntu/anaconda3/lib/python3...
Traceback (most recent call last): File "/home/ubuntu/run_ray_tune.py", line 222, in <module> tune_helsinki_(args) File "/home/ubuntu/run_ray_tune.py", line 106, in tune_helsinki_ ray_wandb_func = wandb_mixin(ray_func) File "/home/ubuntu/anaconda3/lib/python3.7/site-packages/ray/tune/integration/wandb.py", line 142, in...
AttributeError
def Concurrently(ops: List[LocalIterator], *, mode="round_robin", output_indexes=None, round_robin_weights=None): """Operator that runs the given parent iterators concurrently. Args: mode (str): One of 'round_robin', 'async'. In 'round...
def Concurrently(ops: List[LocalIterator], *, mode="round_robin", output_indexes=None, round_robin_weights=None): """Operator that runs the given parent iterators concurrently. Args: mode (str): One of {'round_robin', 'async'}. ...
[{'piece_type': 'other', 'piece_content': 'from ray.tune.logger import pretty_print\\nfrom ray.rllib.agents.dqn.dqn import DEFAULT_CONFIG\\nfrom ray.rllib.agents.dqn.dqn import DQNTrainer\\nimport numpy as np\\n\\nconfig = DEFAULT_CONFIG.copy()\\nconfig["env"] = "CartPole-v0"\\nconfig["input"] = {"demo-out": 0.3, "samp...
Traceback (most recent call last): File "/home/enes/ws/code/arl/mt/test/consume_experiences.py", line 16, in <module> results = trainer.train() File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 522, in train raise e File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray...
KeyError
def ParallelRollouts(workers: WorkerSet, *, mode="bulk_sync", num_async=1) -> LocalIterator[SampleBatch]: """Operator to collect experiences in parallel from rollout workers. If there are no remote workers, experiences will be collected serially from the local worker instance instead. ...
def ParallelRollouts(workers: WorkerSet, *, mode="bulk_sync", num_async=1) -> LocalIterator[SampleBatch]: """Operator to collect experiences in parallel from rollout workers. If there are no remote workers, experiences will be collected serially from the local worker instance instead. ...
[{'piece_type': 'other', 'piece_content': 'from ray.tune.logger import pretty_print\\nfrom ray.rllib.agents.dqn.dqn import DEFAULT_CONFIG\\nfrom ray.rllib.agents.dqn.dqn import DQNTrainer\\nimport numpy as np\\n\\nconfig = DEFAULT_CONFIG.copy()\\nconfig["env"] = "CartPole-v0"\\nconfig["input"] = {"demo-out": 0.3, "samp...
Traceback (most recent call last): File "/home/enes/ws/code/arl/mt/test/consume_experiences.py", line 16, in <module> results = trainer.train() File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 522, in train raise e File "/home/enes/ws/envs/rlws/lib/python3.7/site-packages/ray...
KeyError
def _get_node_specific_docker_config(self, node_id): if "docker" not in self.config: return {} docker_config = copy.deepcopy(self.config.get("docker", {})) node_specific_docker = self._get_node_type_specific_fields( node_id, "docker") docker_config.update(node...
def _get_node_specific_docker_config(self, node_id): docker_config = copy.deepcopy(self.config.get("docker", {})) node_specific_docker = self._get_node_type_specific_fields( node_id, "docker") docker_config.update(node_specific_docker) return docker_config
[{'piece_type': 'other', 'piece_content': 'min_workers: 1\\nmax_workers: 1\\ninitial_workers: 1'}, {'piece_type': 'other', 'piece_content': '(myenv) tcw@pop-os:~/projects/ray$ ray up example-full.yaml\\nCommands running under a login shell can produce more output than special processing can handle.\\nThus, the output f...
==> /tmp/ray/session_2020-09-09_17-20-39_779593_74/logs/monitor.err <== docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/sit...
KeyError
def _clean_log(obj): # Fixes https://github.com/ray-project/ray/issues/10631 if isinstance(obj, dict): return {k: _clean_log(v) for k, v in obj.items()} elif isinstance(obj, list): return [_clean_log(v) for v in obj] # Else try: pickle.dumps(obj) yaml.dump( ...
def _clean_log(obj): # Fixes https://github.com/ray-project/ray/issues/10631 if isinstance(obj, dict): return {k: _clean_log(v) for k, v in obj.items()} elif isinstance(obj, list): return [_clean_log(v) for v in obj] # Else try: pickle.dumps(obj) return obj excep...
[{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi...
Process _WandbLoggingProcess-1: Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap self.run() File "[...]/ray/tune/integration/wandb.py", line 127, in run wandb.init(*self.args, **self.kwargs) File "[...]/wandb/__init__.py", line 1303, in init as_defaults=no...
yaml.representer.RepresenterError
def _init(self): config = self.config.copy() config.pop("callbacks", None) # Remove callbacks try: if config.get("logger_config", {}).get("wandb"): logger_config = config.pop("logger_config") wandb_config = logger_config.get("wandb").copy() ...
def _init(self): config = self.config.copy() try: if config.get("logger_config", {}).get("wandb"): logger_config = config.pop("logger_config") wandb_config = logger_config.get("wandb").copy() else: wandb_config = config.pop("wa...
[{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi...
Process _WandbLoggingProcess-1: Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap self.run() File "[...]/ray/tune/integration/wandb.py", line 127, in run wandb.init(*self.args, **self.kwargs) File "[...]/wandb/__init__.py", line 1303, in init as_defaults=no...
yaml.representer.RepresenterError
def _clean_log(obj): # Fixes https://github.com/ray-project/ray/issues/10631 if isinstance(obj, dict): return {k: _clean_log(v) for k, v in obj.items()} elif isinstance(obj, list): return [_clean_log(v) for v in obj] elif _is_allowed_type(obj): return obj # Else try: ...
def _clean_log(obj): # Fixes https://github.com/ray-project/ray/issues/10631 if isinstance(obj, dict): return {k: _clean_log(v) for k, v in obj.items()} elif isinstance(obj, list): return [_clean_log(v) for v in obj] # Else try: pickle.dumps(obj) yaml.dump( ...
[{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi...
Process _WandbLoggingProcess-1: Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap self.run() File "[...]/ray/tune/integration/wandb.py", line 127, in run wandb.init(*self.args, **self.kwargs) File "[...]/wandb/__init__.py", line 1303, in init as_defaults=no...
yaml.representer.RepresenterError
def _handle_result(self, result): config_update = result.get("config", {}).copy() log = {} flat_result = flatten_dict(result, delimiter="/") for k, v in flat_result.items(): if any( k.startswith(item + "/") or k == item for item in...
def _handle_result(self, result): config_update = result.get("config", {}).copy() log = {} flat_result = flatten_dict(result, delimiter="/") for k, v in flat_result.items(): if any( k.startswith(item + "/") or k == item for item in...
[{'piece_type': 'error message', 'piece_content': 'Process _WandbLoggingProcess-1:\\nTraceback (most recent call last):\\nFile "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap\\nself.run()\\nFile "[...]/ray/tune/integration/wandb.py", line 127, in run\\nwandb.init(*self.args, **self.kwargs)\\nFi...
Process _WandbLoggingProcess-1: Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap self.run() File "[...]/ray/tune/integration/wandb.py", line 127, in run wandb.init(*self.args, **self.kwargs) File "[...]/wandb/__init__.py", line 1303, in init as_defaults=no...
yaml.representer.RepresenterError
def start(node_ip_address, redis_address, address, redis_port, port, num_redis_shards, redis_max_clients, redis_password, redis_shard_ports, object_manager_port, node_manager_port, gcs_server_port, min_worker_port, max_worker_port, memory, object_store_memory, redis_max_memory, n...
def start(node_ip_address, redis_address, address, redis_port, port, num_redis_shards, redis_max_clients, redis_password, redis_shard_ports, object_manager_port, node_manager_port, gcs_server_port, min_worker_port, max_worker_port, memory, object_store_memory, redis_max_memory, n...
[{'piece_type': 'error message', 'piece_content': '$ ray memory\\n2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ubuntu/anaconda3/lib/pytho...
$ ray memory 2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379. Traceback (most recent call last): File "/home/ubuntu/anaconda3/bin/ray", line 8, in <module> sys.exit(main()) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1602, in mai...
TypeError
def memory(address, redis_password): """Print object references held in a Ray cluster.""" if not address: address = services.find_redis_address_or_die() logger.info(f"Connecting to Ray instance at {address}.") ray.init(address=address, _redis_password=redis_password) print(ray.internal.inter...
def memory(address, redis_password): """Print object references held in a Ray cluster.""" if not address: address = services.find_redis_address_or_die() logger.info(f"Connecting to Ray instance at {address}.") ray.init(address=address, redis_password=redis_password) print(ray.internal.intern...
[{'piece_type': 'error message', 'piece_content': '$ ray memory\\n2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379.\\nTraceback (most recent call last):\\nFile "/home/ubuntu/anaconda3/bin/ray", line 8, in <module>\\nsys.exit(main())\\nFile "/home/ubuntu/anaconda3/lib/pytho...
$ ray memory 2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379. Traceback (most recent call last): File "/home/ubuntu/anaconda3/bin/ray", line 8, in <module> sys.exit(main()) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1602, in mai...
TypeError
def choose_trial_to_run(self, trial_runner, allow_recurse=True): """Fair scheduling within iteration by completion percentage. List of trials not used since all trials are tracked as state of scheduler. If iteration is occupied (ie, no trials to run), then look into next iteration. ...
def choose_trial_to_run(self, trial_runner): """Fair scheduling within iteration by completion percentage. List of trials not used since all trials are tracked as state of scheduler. If iteration is occupied (ie, no trials to run), then look into next iteration. """ ...
[{'piece_type': 'error message', 'piece_content': '== Status ==\\nMemory usage on this node: 7.0/15.6 GiB\\nUsing HyperBand: num_stopped=832 total_brackets=3\\nRound #0:\\nNone\\nBracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}\\nBracket(Max Size (n)=324, Milestone (r)=8, com...
== Status == Memory usage on this node: 7.0/15.6 GiB Using HyperBand: num_stopped=832 total_brackets=3 Round #0: None Bracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833} Bracket(Max Size (n)=324, Milestone (r)=8, completed=47.3%): {PAUSED: 166} Resources requested: 4/32 CPUs, 0/...
ray.tune.error.TuneError
def debug_string(self): """This provides a progress notification for the algorithm. For each bracket, the algorithm will output a string as follows: Bracket(Max Size (n)=5, Milestone (r)=33, completed=14.6%): {PENDING: 2, RUNNING: 3, TERMINATED: 2} "Max Size" indic...
def debug_string(self): """This provides a progress notification for the algorithm. For each bracket, the algorithm will output a string as follows: Bracket(Max Size (n)=5, Milestone (r)=33, completed=14.6%): {PENDING: 2, RUNNING: 3, TERMINATED: 2} "Max Size" indic...
[{'piece_type': 'error message', 'piece_content': '== Status ==\\nMemory usage on this node: 7.0/15.6 GiB\\nUsing HyperBand: num_stopped=832 total_brackets=3\\nRound #0:\\nNone\\nBracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833}\\nBracket(Max Size (n)=324, Milestone (r)=8, com...
== Status == Memory usage on this node: 7.0/15.6 GiB Using HyperBand: num_stopped=832 total_brackets=3 Round #0: None Bracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833} Bracket(Max Size (n)=324, Milestone (r)=8, completed=47.3%): {PAUSED: 166} Resources requested: 4/32 CPUs, 0/...
ray.tune.error.TuneError
def run_rsync_up(self, source, target): if target.startswith("~"): target = "/root" + target[1:] try: self.process_runner.check_call([ KUBECTL_RSYNC, "-avz", source, "{}@{}:{}".format(self.node_id, self.namespac...
def run_rsync_up(self, source, target): if target.startswith("~"): target = "/root" + target[1:] try: self.process_runner.check_call([ KUBECTL_RSYNC, "-avz", source, "{}@{}:{}".format(self.node_id, self.namespac...
[{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call...
2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/ Traceback (most recent call last): File "/opt/conda/lib/python3.6/site-packages/ra...
ray.tune.error.TuneError
def run_rsync_down(self, source, target): if target.startswith("~"): target = "/root" + target[1:] try: self.process_runner.check_call([ KUBECTL_RSYNC, "-avz", "{}@{}:{}".format(self.node_id, self.namespace, source), ...
def run_rsync_down(self, source, target): if target.startswith("~"): target = "/root" + target[1:] try: self.process_runner.check_call([ KUBECTL_RSYNC, "-avz", "{}@{}:{}".format(self.node_id, self.namespace, source), ...
[{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call...
2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/ Traceback (most recent call last): File "/opt/conda/lib/python3.6/site-packages/ra...
ray.tune.error.TuneError
def get_node_syncer(local_dir, remote_dir=None, sync_function=None): """Returns a NodeSyncer. Args: local_dir (str): Source directory for syncing. remote_dir (str): Target directory for syncing. If not provided, a noop Syncer is returned. sync_function (func|str|bool): Funct...
def get_node_syncer(local_dir, remote_dir=None, sync_function=None): """Returns a NodeSyncer. Args: local_dir (str): Source directory for syncing. remote_dir (str): Target directory for syncing. If not provided, a noop Syncer is returned. sync_function (func|str|bool): Funct...
[{'piece_type': 'error message', 'piece_content': '2020-07-17 21:53:48,101\\tERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/\\nTraceback (most recent call...
2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/ Traceback (most recent call last): File "/opt/conda/lib/python3.6/site-packages/ra...
ray.tune.error.TuneError
def action_prob(self, batch: SampleBatchType) -> np.ndarray: """Returns the probs for the batch actions for the current policy.""" num_state_inputs = 0 for k in batch.keys(): if k.startswith("state_in_"): num_state_inputs += 1 state_keys = ["state_in_{}"....
def action_prob(self, batch: SampleBatchType) -> TensorType: """Returns the probs for the batch actions for the current policy.""" num_state_inputs = 0 for k in batch.keys(): if k.startswith("state_in_"): num_state_inputs += 1 state_keys = ["state_in_{}"....
[{'piece_type': 'error message', 'piece_content': 'Traceback (most recent call last):\\nFile "C:\\\\Users\\\\Julius\\\\Anaconda3\\\\envs\\\\ray\\\\lib\\\\site-packages\\\\ray\\\\tune\\\\trial_runner.py", line 497, in _process_trial\\nresult = self.trial_executor.fetch_result(trial)\\nFile "C:\\\\Users\\\\Julius\\\\Anac...
Traceback (most recent call last): File "C:\\Users\\Julius\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\tune\\trial_runner.py", line 497, in _process_trial result = self.trial_executor.fetch_result(trial) File "C:\\Users\\Julius\\Anaconda3\\envs\\ray\\lib\\site-packages\\ray\\tune\\ray_trial_executor.py", line 434, ...
AttributeError
def run_rsync_up(self, source, target): # TODO(ilr) Expose this to before NodeUpdater::sync_file_mounts protected_path = target if target.find("/root") == 0: target = target.replace("/root", "/tmp/root") self.ssh_command_runner.run( f"mkdir -p {os.path.dirname...
def run_rsync_up(self, source, target): protected_path = target if target.find("/root") == 0: target = target.replace("/root", "/tmp/root") self.ssh_command_runner.run( f"mkdir -p {os.path.dirname(target.rstrip('/'))}") self.ssh_command_runner.run_rsync_up(sou...
[{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p...
(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml 2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1 2020-08-12 20:12:39,612 INFO c...
AssertionError
def sync_file_mounts(self, sync_cmd, step_numbers=(0, 2)): # step_numbers is (# of previous steps, total steps) previous_steps, total_steps = step_numbers nolog_paths = [] if cli_logger.verbosity == 0: nolog_paths = [ "~/ray_bootstrap_key.pem", "~/ray_boo...
def sync_file_mounts(self, sync_cmd, step_numbers=(0, 2)): # step_numbers is (# of previous steps, total steps) previous_steps, total_steps = step_numbers nolog_paths = [] if cli_logger.verbosity == 0: nolog_paths = [ "~/ray_bootstrap_key.pem", "~/ray_boo...
[{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p...
(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml 2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1 2020-08-12 20:12:39,612 INFO c...
AssertionError
def do_sync(remote_path, local_path, allow_non_existing_paths=False): if allow_non_existing_paths and not os.path.exists(local_path): # Ignore missing source files. In the future we should support # the --delete-missing-args command to delete files that have ...
def do_sync(remote_path, local_path, allow_non_existing_paths=False): if allow_non_existing_paths and not os.path.exists(local_path): # Ignore missing source files. In the future we should support # the --delete-missing-args command to delete files that have ...
[{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p...
(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml 2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1 2020-08-12 20:12:39,612 INFO c...
AssertionError
def wait_ready(self, deadline): with cli_logger.group( "Waiting for SSH to become available", _numbered=("[]", 1, 6)): with LogTimer(self.log_prefix + "Got remote shell"): cli_logger.old_info(logger, "{}Waiting for remote shell...", ...
def wait_ready(self, deadline): with cli_logger.group( "Waiting for SSH to become available", _numbered=("[]", 1, 6)): with LogTimer(self.log_prefix + "Got remote shell"): cli_logger.old_info(logger, "{}Waiting for remote shell...", ...
[{'piece_type': 'error message', 'piece_content': '(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml\\n2020-08-12 20:12:39,383\\tINFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-p...
(vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml 2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1 2020-08-12 20:12:39,612 INFO c...
AssertionError
def create_or_update_cluster(config_file: str, override_min_workers: Optional[int], override_max_workers: Optional[int], no_restart: bool, restart_only: bool, yes: bool, ...
def create_or_update_cluster( config_file: str, override_min_workers: Optional[int], override_max_workers: Optional[int], no_restart: bool, restart_only: bool, yes: bool, override_cluster_name: Optional[str], no_config_cache: bool, dump_command_output: bool, use_login_shells: boo...
[{'piece_type': 'error message', 'piece_content': '(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py\\nTraceback (most recent call last):\\nFile "/Users/alex/miniconda3/bin/ray", line 11, in <module>\\nload_entry_point(\\'ray\\', \\'console_scripts\\', \\'ray\\')()\\nFile "/Users/alex/anyscale/...
(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py Traceback (most recent call last): File "/Users/alex/miniconda3/bin/ray", line 11, in <module> load_entry_point('ray', 'console_scripts', 'ray')() File "/Users/alex/anyscale/ray/python/ray/scripts/scripts.py", line 1587, in main return cli() Fil...
TypeError
def submit(cluster_config_file, screen, tmux, stop, start, cluster_name, port_forward, script, args, script_args, log_new_style, log_color, verbose): """Uploads and runs a script on the specified cluster. The script is automatically synced to the following location: os.path.join(...
def submit(cluster_config_file, screen, tmux, stop, start, cluster_name, port_forward, script, args, script_args, log_new_style, log_color, verbose): """Uploads and runs a script on the specified cluster. The script is automatically synced to the following location: os.path.join(...
[{'piece_type': 'error message', 'piece_content': '(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py\\nTraceback (most recent call last):\\nFile "/Users/alex/miniconda3/bin/ray", line 11, in <module>\\nload_entry_point(\\'ray\\', \\'console_scripts\\', \\'ray\\')()\\nFile "/Users/alex/anyscale/...
(base) Alexs-MacBook-Pro-2:ray alex$ ray submit --start multi.yaml test.py Traceback (most recent call last): File "/Users/alex/miniconda3/bin/ray", line 11, in <module> load_entry_point('ray', 'console_scripts', 'ray')() File "/Users/alex/anyscale/ray/python/ray/scripts/scripts.py", line 1587, in main return cli() Fil...
TypeError