| | |
| | |
| | |
| | |
| |
|
| | import glob |
| | import json |
| | import math |
| | import os |
| | import re |
| | from datetime import datetime |
| |
|
| | import numpy as np |
| | import yaml |
| | from tensorboard.backend.event_processing import event_accumulator |
| |
|
| | import carb |
| |
|
| |
|
| | def get_env_configs(configs_path): |
| | """Get environment configurations from yaml filepath.""" |
| | with open(configs_path) as env_configs_file: |
| | env_configs = yaml.safe_load(env_configs_file) |
| | return env_configs |
| |
|
| |
|
| | def get_env_config(env_configs, mode, workflow, task): |
| | """Get the environment configuration.""" |
| | if mode not in env_configs: |
| | raise ValueError(f"Mode {mode} is not supported in the config file.") |
| |
|
| | extended_task = f"{workflow}:{task}" |
| | |
| | if extended_task in env_configs[mode]: |
| | return env_configs[mode][extended_task] |
| |
|
| | |
| | if task in env_configs[mode]: |
| | return env_configs[mode][task] |
| |
|
| | |
| | for env_config_key in env_configs[mode].keys(): |
| | if re.match(env_config_key, extended_task): |
| | return env_configs[mode][env_config_key] |
| |
|
| | |
| | for env_config_key in env_configs[mode].keys(): |
| | if re.match(env_config_key, task): |
| | return env_configs[mode][env_config_key] |
| |
|
| | |
| | return None |
| |
|
| |
|
| | def evaluate_job(workflow, task, env_config, duration): |
| | """Evaluate the job.""" |
| | log_data = _retrieve_logs(workflow, task) |
| |
|
| | kpi_payload = {"success": True, "msg": ""} |
| |
|
| | |
| | if not log_data: |
| | kpi_payload["success"] = False |
| | kpi_payload["msg"] = "error: training did not finish!" |
| | return kpi_payload |
| |
|
| | thresholds = {**env_config.get("lower_thresholds", {}), **env_config.get("upper_thresholds", {})} |
| |
|
| | |
| | for threshold_name, threshold_val in thresholds.items(): |
| | uses_lower_threshold = threshold_name in env_config.get("lower_thresholds", {}) |
| | if threshold_name == "duration": |
| | val = duration |
| | else: |
| | val = _extract_log_val(threshold_name, log_data, uses_lower_threshold, workflow) |
| | |
| | if val is None or not isinstance(val, (int, float)) or (isinstance(val, float) and math.isnan(val)): |
| | continue |
| | val = round(val, 4) |
| | if uses_lower_threshold: |
| | |
| | if val < threshold_val: |
| | kpi_payload["success"] = False |
| | else: |
| | |
| | if val > threshold_val: |
| | kpi_payload["success"] = False |
| | kpi_payload[threshold_name] = val |
| | if threshold_name == "reward": |
| | normalized_reward = val / threshold_val |
| | kpi_payload[f"{threshold_name}_normalized"] = normalized_reward |
| | kpi_payload[f"{threshold_name}_threshold"] = threshold_val |
| |
|
| | |
| | max_iterations = env_config.get("max_iterations") |
| | if max_iterations is not None: |
| | kpi_payload["max_iterations"] = max_iterations |
| |
|
| | return kpi_payload |
| |
|
| |
|
| | def process_kpi_data(kpi_payloads, tag=""): |
| | """Combine and augment the KPI payloads.""" |
| | |
| | totals = {} |
| | successes = {} |
| | failures_did_not_finish = {} |
| | failures_did_not_pass_thresholds = {} |
| | for job_id, kpi_payload in kpi_payloads.items(): |
| | workflow = job_id.split(":")[0] |
| | if workflow not in totals: |
| | totals[workflow] = 0 |
| | successes[workflow] = 0 |
| | failures_did_not_finish[workflow] = 0 |
| | failures_did_not_pass_thresholds[workflow] = 0 |
| | totals[workflow] += 1 |
| | if kpi_payload["success"]: |
| | successes[workflow] += 1 |
| | else: |
| | if kpi_payload["msg"] == "error: training did not finish!": |
| | failures_did_not_finish[workflow] += 1 |
| | else: |
| | failures_did_not_pass_thresholds[workflow] += 1 |
| |
|
| | kpi_payloads["overall"] = { |
| | "totals": totals, |
| | "successes": successes, |
| | "failures_did_not_finish": failures_did_not_finish, |
| | "failures_did_not_pass_thresholds": failures_did_not_pass_thresholds, |
| | "timestamp": datetime.now().isoformat(), |
| | "tag": tag, |
| | } |
| |
|
| | return kpi_payloads |
| |
|
| |
|
| | def output_payloads(payloads): |
| | """Output the KPI payloads to a json file.""" |
| | |
| | repo_path = os.path.join(carb.tokens.get_tokens_interface().resolve("${app}"), "..") |
| | output_path = os.path.join(repo_path, "logs/kpi.json") |
| | |
| | if not os.path.exists(os.path.dirname(output_path)): |
| | os.makedirs(os.path.dirname(output_path)) |
| | |
| | with open(output_path, "w") as payload_file: |
| | json.dump(payloads, payload_file, indent=4) |
| |
|
| |
|
| | def _retrieve_logs(workflow, task): |
| | """Retrieve training logs.""" |
| | |
| | repo_path = os.path.join(carb.tokens.get_tokens_interface().resolve("${app}"), "..") |
| | from isaaclab.utils.version import get_isaac_sim_version |
| |
|
| | if get_isaac_sim_version().major < 5: |
| | repo_path = os.path.join(repo_path, "..") |
| | if workflow == "rl_games": |
| | log_files_path = os.path.join(repo_path, f"logs/{workflow}/{task}/*/summaries/*") |
| | else: |
| | log_files_path = os.path.join(repo_path, f"logs/{workflow}/{task}/*/*.tfevents.*") |
| | log_files = glob.glob(log_files_path) |
| | |
| | if not log_files: |
| | return None |
| | |
| | latest_log_file = max(log_files, key=os.path.getctime) |
| | |
| | log_data = _parse_tf_logs(latest_log_file) |
| | return log_data |
| |
|
| |
|
| | def _parse_tf_logs(log): |
| | """Parse the tensorflow filepath into a dictionary.""" |
| | log_data = {} |
| | ea = event_accumulator.EventAccumulator(log) |
| | ea.Reload() |
| | tags = ea.Tags()["scalars"] |
| | for tag in tags: |
| | log_data[tag] = [] |
| | for event in ea.Scalars(tag): |
| | log_data[tag].append((event.step, event.value)) |
| | return log_data |
| |
|
| |
|
| | def _extract_log_val(name, log_data, uses_lower_threshold, workflow): |
| | """Extract the value from the log data.""" |
| | try: |
| | if name == "reward": |
| | reward_tags = { |
| | "rl_games": "rewards/iter", |
| | "rsl_rl": "Train/mean_reward", |
| | "sb3": None, |
| | "skrl": "Reward / Total reward (mean)", |
| | } |
| | tag = reward_tags.get(workflow) |
| | if tag: |
| | return _extract_reward(log_data, tag) |
| |
|
| | elif name == "episode_length": |
| | episode_tags = { |
| | "rl_games": "episode_lengths/iter", |
| | "rsl_rl": "Train/mean_episode_length", |
| | "sb3": None, |
| | "skrl": "Episode / Total timesteps (mean)", |
| | } |
| | tag = episode_tags.get(workflow) |
| | if tag: |
| | return _extract_feature(log_data, tag, uses_lower_threshold) |
| |
|
| | elif name == "training_time": |
| | return {"rl_games": log_data["rewards/time"][-1][0], "rsl_rl": None, "sb3": None, "skrl": None}.get( |
| | workflow |
| | ) |
| | except Exception: |
| | return None |
| |
|
| | raise ValueError(f"Env Config name {name} is not supported.") |
| |
|
| |
|
| | def _extract_feature(log_data, feature, uses_lower_threshold): |
| | """Extract the feature from the log data.""" |
| | log_data = np.array(log_data[feature])[:, 1] |
| |
|
| | if uses_lower_threshold: |
| | return max(log_data) |
| | else: |
| | return min(log_data) |
| |
|
| |
|
| | def _extract_reward(log_data, feature, k=8): |
| | """Extract the averaged max reward from the log data.""" |
| | log_data = np.array(log_data[feature])[:, 1] |
| |
|
| | |
| | k = min(len(log_data), k) |
| | averaged_reward = np.mean(np.partition(log_data, -k)[-k:]) |
| |
|
| | return averaged_reward |
| |
|