file_path
stringlengths
3
280
file_language
stringclasses
66 values
content
stringlengths
1
1.04M
repo_name
stringlengths
5
92
repo_stars
int64
0
154k
repo_description
stringlengths
0
402
repo_primary_language
stringclasses
108 values
developer_username
stringlengths
1
25
developer_name
stringlengths
0
30
developer_company
stringlengths
0
82
rllib/agents/ppo/ppo_policy.py
Python
import logging import ray from ray.rllib.evaluation.postprocessing import compute_advantages, \ Postprocessing from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.policy.tf_policy import LearningRateSchedule, \ EntropyCoeffSchedule, ACTION_LOGP from ray.rllib.policy.tf_policy_template import b...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/ppo/test/test.py
Python
import unittest import numpy as np from numpy.testing import assert_allclose from ray.rllib.models.tf.tf_action_dist import Categorical from ray.rllib.agents.ppo.utils import flatten, concatenate from ray.rllib.utils import try_import_tf tf = try_import_tf() # TODO(ekl): move to rllib/models dir class Distributions...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/ppo/utils.py
Python
import numpy as np def flatten(weights, start=0, stop=2): """This methods reshapes all values in a dictionary. The indices from start to stop will be flattened into a single index. Args: weights: A dictionary mapping keys to numpy arrays. start: The starting index. stop: The endi...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/__init__.py
Python
from ray.rllib.agents.qmix.qmix import QMixTrainer, DEFAULT_CONFIG from ray.rllib.agents.qmix.apex import ApexQMixTrainer __all__ = ["QMixTrainer", "ApexQMixTrainer", "DEFAULT_CONFIG"]
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/apex.py
Python
"""Experimental: scalable Ape-X variant of QMIX""" from ray.rllib.agents.dqn.apex import APEX_TRAINER_PROPERTIES from ray.rllib.agents.qmix.qmix import QMixTrainer, \ DEFAULT_CONFIG as QMIX_CONFIG from ray.rllib.utils import merge_dicts APEX_QMIX_DEFAULT_CONFIG = merge_dicts( QMIX_CONFIG, # see also the opti...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/mixers.py
Python
import numpy as np from ray.rllib.utils.framework import try_import_torch torch, nn = try_import_torch() F = nn.functional class VDNMixer(nn.Module): def __init__(self): super(VDNMixer, self).__init__() def forward(self, agent_qs, batch): return torch.sum(agent_qs, dim=2, keepdim=True) cl...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/model.py
Python
from ray.rllib.models.preprocessors import get_preprocessor from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_torch torch, nn = try_import_torch() F = nn.functional class RNNModel(TorchModelV2, nn.Module): """The ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/qmix.py
Python
from ray.rllib.agents.trainer import with_common_config from ray.rllib.agents.dqn.dqn import GenericOffPolicyTrainer from ray.rllib.agents.qmix.qmix_policy import QMixTorchPolicy from ray.rllib.optimizers import SyncBatchReplayOptimizer # yapf: disable # __sphinx_doc_begin__ DEFAULT_CONFIG = with_common_config({ #...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/qmix/qmix_policy.py
Python
from gym.spaces import Tuple, Discrete, Dict import logging import numpy as np import torch as th import torch.nn as nn from torch.optim import RMSprop from torch.distributions import Categorical import ray from ray.rllib.agents.qmix.mixers import VDNMixer, QMixer from ray.rllib.agents.qmix.model import RNNModel, _get...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/registry.py
Python
"""Registry of algorithm names for `rllib train --run=<alg_name>`""" import traceback from ray.rllib.contrib.registry import CONTRIBUTED_ALGORITHMS def _import_sac(): from ray.rllib.agents import sac return sac.SACTrainer def _import_appo(): from ray.rllib.agents import ppo return ppo.APPOTrainer ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/sac/__init__.py
Python
from ray.rllib.agents.sac.sac import SACTrainer, DEFAULT_CONFIG from ray.rllib.utils import renamed_agent SACAgent = renamed_agent(SACTrainer) __all__ = [ "SACTrainer", "DEFAULT_CONFIG", ]
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/sac/sac.py
Python
from ray.rllib.agents.trainer import with_common_config from ray.rllib.agents.dqn.dqn import GenericOffPolicyTrainer from ray.rllib.agents.sac.sac_policy import SACTFPolicy OPTIMIZER_SHARED_CONFIGS = [ "buffer_size", "prioritized_replay", "prioritized_replay_alpha", "prioritized_replay_beta", "prioritized_repl...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/sac/sac_model.py
Python
import numpy as np from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.utils import try_import_tf, try_import_tfp tf = try_import_tf() tfp = try_import_tfp() SCALE_DIAG_MIN_MAX = (-20, 2) def SquashBijector(): # lazy def since it depends on tfp class SquashBijector(tfp.bijectors.Bijector): ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/sac/sac_policy.py
Python
from gym.spaces import Box import numpy as np import logging import ray import ray.experimental.tf_utils from ray.rllib.agents.sac.sac_model import SACModel from ray.rllib.agents.ddpg.noop_model import NoopModel from ray.rllib.agents.dqn.dqn_policy import _postprocess_dqn, PRIO_WEIGHTS from ray.rllib.policy.sample_bat...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/trainer.py
Python
from datetime import datetime import copy import logging import os import pickle import six import time import tempfile import ray from ray.exceptions import RayError from ray.rllib.models import MODEL_DEFAULTS from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID from ray.rllib.evaluation.metrics import collect...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/agents/trainer_template.py
Python
import time from ray.rllib.agents.trainer import Trainer, COMMON_CONFIG from ray.rllib.optimizers import SyncSamplesOptimizer from ray.rllib.utils import add_mixins from ray.rllib.utils.annotations import override, DeveloperAPI @DeveloperAPI def build_trainer(name, default_policy, ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/core/alpha_zero_policy.py
Python
import numpy as np from ray.rllib.policy.policy import Policy, LEARNER_STATS_KEY from ray.rllib.policy.torch_policy import TorchPolicy from ray.rllib.utils.annotations import override from ray.rllib.contrib.alpha_zero.core.mcts import Node, RootParentNode from ray.rllib.utils import try_import_torch torch, _ = try_im...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/core/alpha_zero_trainer.py
Python
import logging from ray.rllib.agents import with_common_config from ray.rllib.agents.trainer_template import build_trainer from ray.rllib.models.catalog import ModelCatalog from ray.rllib.models.model import restore_original_dimensions from ray.rllib.models.torch.torch_action_dist import TorchCategorical from ray.rlli...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/core/mcts.py
Python
""" Mcts implementation modified from https://github.com/brilee/python_uct/blob/master/numpy_impl.py """ import collections import math import numpy as np class Node: def __init__(self, action, obs, done, reward, state, mcts, parent=None): self.env = parent.env self.action = action # Action used...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/core/ranked_rewards.py
Python
from copy import deepcopy import numpy as np class RankedRewardsBuffer: def __init__(self, buffer_max_length, percentile): self.buffer_max_length = buffer_max_length self.percentile = percentile self.buffer = [] def add_reward(self, reward): if len(self.buffer) < self.buffer_...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/environments/cartpole.py
Python
from copy import deepcopy import gym import numpy as np from gym.spaces import Discrete, Dict, Box class CartPole: """ Wrapper for gym CartPole environment where the reward is accumulated to the end """ def __init__(self, config=None): self.env = gym.make("CartPole-v0") self.acti...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/examples/train_cartpole.py
Python
"""Example of using training on CartPole.""" import argparse from ray import tune from ray.rllib.contrib.alpha_zero.models.custom_torch_models import DenseModel from ray.rllib.contrib.alpha_zero.environments.cartpole import CartPole from ray.rllib.models.catalog import ModelCatalog if __name__ == "__main__": pa...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/models/custom_torch_models.py
Python
from abc import ABC import numpy as np from ray.rllib.models.model import restore_original_dimensions from ray.rllib.models.preprocessors import get_preprocessor from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.utils import try_import_torch torch, nn = try_import_torch() def convert_to_...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/alpha_zero/optimizer/sync_batches_replay_optimizer.py
Python
import random from ray.rllib.evaluation.metrics import get_learner_stats from ray.rllib.optimizers.sync_batch_replay_optimizer import \ SyncBatchReplayOptimizer from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.utils.annotations import override class SyncBatchesReplayOptimizer(SyncBatchReplayO...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/maddpg/__init__.py
Python
from ray.rllib.contrib.maddpg.maddpg import MADDPGTrainer, DEFAULT_CONFIG __all__ = ["MADDPGTrainer", "DEFAULT_CONFIG"]
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/maddpg/maddpg.py
Python
"""Contributed port of MADDPG from OpenAI baselines. The implementation has a couple assumptions: - The number of agents is fixed and known upfront. - Each agent is bound to a policy of the same name. - Discrete actions are sent as logits (pre-softmax). For a minimal example, see twostep_game.py, and the README for h...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/maddpg/maddpg_policy.py
Python
import ray from ray.rllib.agents.dqn.dqn_policy import minimize_and_clip, _adjust_nstep from ray.rllib.evaluation.metrics import LEARNER_STATS_KEY from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.models import ModelCatalog from ray.rllib.utils.annotations import override from ray.rllib.utils.error i...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/random_agent/random_agent.py
Python
import numpy as np from ray.rllib.agents.trainer import Trainer, with_common_config from ray.rllib.utils.annotations import override # yapf: disable # __sphinx_doc_begin__ class RandomAgent(Trainer): """Policy that takes random actions and never learns.""" _name = "RandomAgent" _default_config = with_co...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/contrib/registry.py
Python
"""Registry of algorithm names for `rllib train --run=<alg_name>`""" def _import_random_agent(): from ray.rllib.contrib.random_agent.random_agent import RandomAgent return RandomAgent def _import_maddpg(): from ray.rllib.contrib import maddpg return maddpg.MADDPGTrainer def _import_alphazero(): ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/__init__.py
Python
from ray.rllib.evaluation.episode import MultiAgentEpisode from ray.rllib.evaluation.rollout_worker import RolloutWorker from ray.rllib.evaluation.policy_evaluator import PolicyEvaluator from ray.rllib.evaluation.interface import EvaluatorInterface from ray.rllib.evaluation.policy_graph import PolicyGraph from ray.rlli...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/episode.py
Python
from collections import defaultdict import random import numpy as np from ray.rllib.env.base_env import _DUMMY_AGENT_ID from ray.rllib.utils.annotations import DeveloperAPI @DeveloperAPI class MultiAgentEpisode: """Tracks the current state of a (possibly multi-agent) episode. Attributes: new_batch_...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/interface.py
Python
import os from ray.rllib.utils.annotations import DeveloperAPI @DeveloperAPI class EvaluatorInterface: """This is the interface between policy optimizers and policy evaluation. See also: RolloutWorker """ @DeveloperAPI def sample(self): """Returns a batch of experience sampled from this...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/metrics.py
Python
import logging import numpy as np import collections import ray from ray.rllib.evaluation.rollout_metrics import RolloutMetrics from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID from ray.rllib.offline.off_policy_estimator import OffPolicyEstimate from ray.rllib.policy.policy import LEARNER_STATS_KEY from ray...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/policy_evaluator.py
Python
from ray.rllib.utils import renamed_class from ray.rllib.evaluation import RolloutWorker PolicyEvaluator = renamed_class( RolloutWorker, old_name="rllib.evaluation.PolicyEvaluator")
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/policy_graph.py
Python
from ray.rllib.policy.policy import Policy from ray.rllib.utils import renamed_class PolicyGraph = renamed_class(Policy, old_name="PolicyGraph")
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/postprocessing.py
Python
import numpy as np import scipy.signal from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.utils.annotations import DeveloperAPI def discount(x, gamma): return scipy.signal.lfilter([1], [1, -gamma], x[::-1], axis=0)[::-1] class Postprocessing: """Constant definitions for postprocessing.""" ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/rollout_metrics.py
Python
import collections # Define this in its own file, see #5125 RolloutMetrics = collections.namedtuple("RolloutMetrics", [ "episode_length", "episode_reward", "agent_rewards", "custom_metrics", "perf_stats" ])
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/rollout_worker.py
Python
import random import numpy as np import gym import logging import pickle import ray from ray.rllib.env.atari_wrappers import wrap_deepmind, is_atari from ray.rllib.env.base_env import BaseEnv from ray.rllib.env.env_context import EnvContext from ray.rllib.env.external_env import ExternalEnv from ray.rllib.env.multi_ag...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/sample_batch.py
Python
from ray.rllib.policy.sample_batch import SampleBatch, MultiAgentBatch from ray.rllib.utils import renamed_class SampleBatch = renamed_class( SampleBatch, old_name="rllib.evaluation.SampleBatch") MultiAgentBatch = renamed_class( MultiAgentBatch, old_name="rllib.evaluation.MultiAgentBatch")
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/sample_batch_builder.py
Python
import collections import logging import numpy as np from ray.rllib.policy.sample_batch import SampleBatch, MultiAgentBatch from ray.rllib.utils.annotations import PublicAPI, DeveloperAPI from ray.rllib.utils.debug import log_once, summarize logger = logging.getLogger(__name__) def to_float_array(v): arr = np.a...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/sampler.py
Python
from collections import defaultdict, namedtuple import logging import numpy as np import six.moves.queue as queue import threading import time from ray.rllib.evaluation.episode import MultiAgentEpisode, _flatten_action from ray.rllib.evaluation.rollout_metrics import RolloutMetrics from ray.rllib.evaluation.sample_bat...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/tf_policy_graph.py
Python
from ray.rllib.policy.tf_policy import TFPolicy from ray.rllib.utils import renamed_class TFPolicyGraph = renamed_class(TFPolicy, old_name="TFPolicyGraph")
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/torch_policy_graph.py
Python
from ray.rllib.policy.torch_policy import TorchPolicy from ray.rllib.utils import renamed_class TorchPolicyGraph = renamed_class(TorchPolicy, old_name="TorchPolicyGraph")
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/evaluation/worker_set.py
Python
import logging from types import FunctionType from ray.rllib.utils.annotations import DeveloperAPI from ray.rllib.evaluation.rollout_worker import RolloutWorker, \ _validate_multiagent_config from ray.rllib.offline import NoopOutput, JsonReader, MixedInput, JsonWriter, \ ShuffledInput from ray.rllib.utils impo...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/autoregressive_action_dist.py
Python
"""Example of specifying an autoregressive action distribution. In an action space with multiple components (e.g., Tuple(a1, a2)), you might want a2 to be sampled based on the sampled value of a1, i.e., a2_sampled ~ P(a2 | a1_sampled, obs). Normally, a1 and a2 would be sampled independently. To do this, you need both...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/batch_norm_model.py
Python
"""Example of using a custom model with batch norm.""" import argparse import ray from ray import tune from ray.rllib.models import Model, ModelCatalog from ray.rllib.models.tf.misc import normc_initializer from ray.rllib.utils import try_import_tf tf = try_import_tf() parser = argparse.ArgumentParser() parser.add_...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/cartpole_lstm.py
Python
"""Partially observed variant of the CartPole gym environment. https://github.com/openai/gym/blob/master/gym/envs/classic_control/cartpole.py We delete the velocity component of the state, so that it can only be solved by a LSTM policy.""" import argparse import math import gym from gym import spaces from gym.utils ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/centralized_critic.py
Python
"""An example of customizing PPO to leverage a centralized critic. Here the model and policy are hard-coded to implement a centralized critic for TwoStepGame, but you can adapt this for your own use cases. Compared to simply running `twostep_game.py --run=PPO`, this centralized critic version reaches vf_explained_var...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/centralized_critic_2.py
Python
"""An example of implementing a centralized critic by modifying the env. The advantage of this approach is that it's very simple and you don't have to change the algorithm at all -- just use an env wrapper and custom model. However, it is a bit less principled in that you have to change the agent observation spaces an...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_env.py
Python
"""Example of a custom gym environment and model. Run this for a demo. This example shows: - using a custom environment - using a custom model - using Tune for grid search You can visualize experiment results in ~/ray_results using TensorBoard. """ import numpy as np import gym from ray.rllib.models import Mod...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_fast_model.py
Python
"""Example of using a custom image env and model. Both the model and env are trivial (and super-fast), so they are useful for running perf microbenchmarks. """ from gym.spaces import Discrete, Box import gym import numpy as np import ray from ray.rllib.models import Model, ModelCatalog from ray.tune import run_exper...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_keras_model.py
Python
"""Example of using a custom ModelV2 Keras-style model.""" import argparse import ray from ray import tune from ray.rllib.models import ModelCatalog from ray.rllib.models.tf.misc import normc_initializer from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.agents.dqn.distributional_q_model import Distr...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_keras_rnn_model.py
Python
"""Example of using a custom RNN keras model.""" import gym from gym.spaces import Discrete import numpy as np import random import argparse import ray from ray import tune from ray.tune.registry import register_env from ray.rllib.models import ModelCatalog from ray.rllib.models.modelv2 import ModelV2 from ray.rllib....
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_loss.py
Python
"""Example of using custom_loss() with an imitation learning loss. The default input file is too small to learn a good policy, but you can generate new experiences for IL training as follows: To generate experiences: $ ./train.py --run=PG --config='{"output": "/tmp/cartpole"}' --env=CartPole-v0 To train on experienc...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_metrics_and_callbacks.py
Python
"""Example of using RLlib's debug callbacks. Here we use callbacks to track the average CartPole pole angle magnitude as a custom metric. """ import argparse import numpy as np import ray from ray import tune def on_episode_start(info): episode = info["episode"] print("episode {} started".format(episode.ep...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_tf_policy.py
Python
import argparse import ray from ray import tune from ray.rllib.agents.trainer_template import build_trainer from ray.rllib.evaluation.postprocessing import discount from ray.rllib.policy.tf_policy_template import build_tf_policy from ray.rllib.utils import try_import_tf tf = try_import_tf() parser = argparse.Argumen...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_torch_policy.py
Python
import argparse import ray from ray import tune from ray.rllib.agents.trainer_template import build_trainer from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.policy.torch_policy_template import build_torch_policy parser = argparse.ArgumentParser() parser.add_argument("--iters", type=int, default=20...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/custom_train_fn.py
Python
"""Example of a custom training workflow. Run this for a demo. This example shows: - using Tune trainable functions to implement custom training workflows You can visualize experiment results in ~/ray_results using TensorBoard. """ import ray from ray import tune from ray.rllib.agents.ppo import PPOTrainer def m...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/dmlab_watermaze.py
Python
from deepmind_lab import dmenv_module from ray.rllib import env class Watermaze(env.DMEnv): def __init__(self, env_config): lab = dmenv_module.Lab( "contributed/dmlab30/rooms_watermaze", ["RGBD"], config=env_config, ) super(Watermaze, self).__init__(lab...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/eager_execution.py
Python
import argparse import random import ray from ray import tune from ray.rllib.agents.trainer_template import build_trainer from ray.rllib.models import Model, ModelCatalog from ray.rllib.models.tf.fcnet_v1 import FullyConnectedNetwork from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.policy.tf_policy...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/export/cartpole_dqn_export.py
Python
#!/usr/bin/env python import os import ray from ray.rllib.agents.registry import get_agent_class from ray.rllib.utils import try_import_tf tf = try_import_tf() ray.init(num_cpus=10) def train_and_export(algo_name, num_steps, model_dir, ckpt_dir, prefix): cls = get_agent_class(algo_name) alg = cls(config={...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/hierarchical_training.py
Python
"""Example of hierarchical training using the multi-agent API. The example env is that of a "windy maze". The agent observes the current wind direction and can either choose to stand still, or move in that direction. You can try out the env directly with: $ python hierarchical_training.py --flat A simple hierar...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/multiagent_cartpole.py
Python
"""Simple example of setting up a multi-agent policy mapping. Control the number of agents and policies via --num-agents and --num-policies. This works with hundreds of agents and policies, but note that initializing many TF policies will take some time. Also, TF evals might slow down with large numbers of policies....
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/multiagent_custom_policy.py
Python
"""Example of running a custom hand-coded policy alongside trainable policies. This example has two policies: (1) a simple PG policy (2) a hand-coded policy that acts at random in the env (doesn't learn) In the console output, you can see the PG policy does much better than random: Result for PG_multi_cartpol...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/multiagent_two_trainers.py
Python
"""Example of using two different training methods at once in multi-agent. Here we create a number of CartPole agents, some of which are trained with DQN, and some of which are trained with PPO. We periodically sync weights between the two trainers (note that no such syncing is needed when using just a single training...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/parametric_action_cartpole.py
Python
"""Example of handling variable length and/or parametric action spaces. This is a toy example of the action-embedding based approach for handling large discrete action spaces (potentially infinite in size), similar to this: https://neuro.cs.ut.ee/the-use-of-embeddings-in-openai-five/ This currently works with RL...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/rock_paper_scissors_multiagent.py
Python
"""A simple multi-agent env with two agents playing rock paper scissors. This demonstrates running the following policies in competition: (1) heuristic policy of repeating the same move (2) heuristic policy of beating the last opponent move (3) LSTM/feedforward PG policies (4) LSTM policy with custom e...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/rollout_worker_custom_workflow.py
Python
"""Example of using rollout worker classes directly to implement training. Instead of using the built-in Trainer classes provided by RLlib, here we define a custom Policy class and manually coordinate distributed sample collection and policy optimization. """ import argparse import gym import ray from ray import tun...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/saving_experiences.py
Python
"""Simple example of writing experiences to a file using JsonWriter.""" # __sphinx_doc_begin__ import gym import numpy as np from ray.rllib.models.preprocessors import get_preprocessor from ray.rllib.evaluation.sample_batch_builder import SampleBatchBuilder from ray.rllib.offline.json_writer import JsonWriter if __n...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/serving/cartpole_client.py
Python
"""Example of querying a policy server. Copy this file for your use case. To try this out, in two separate shells run: $ python cartpole_server.py $ python cartpole_client.py """ import argparse import gym from ray.rllib.utils.policy_client import PolicyClient parser = argparse.ArgumentParser() parser.add_a...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/serving/cartpole_server.py
Python
"""Example of running a policy server. Copy this file for your use case. To try this out, in two separate shells run: $ python cartpole_server.py $ python cartpole_client.py """ import os from gym import spaces import numpy as np import ray from ray.rllib.agents.dqn import DQNTrainer from ray.rllib.env.exter...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/serving/test.sh
Shell
#!/bin/bash pkill -f cartpole_server.py (python cartpole_server.py 2>&1 | grep -v 200) & pid=$! while ! curl localhost:9900; do sleep 1 done python cartpole_client.py --stop-at-reward=100 kill $pid
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/examples/twostep_game.py
Python
"""The two-step game from QMIX: https://arxiv.org/pdf/1803.11485.pdf Configurations you can try: - normal policy gradients (PG) - contrib/MADDPG - QMIX - APEX_QMIX See also: centralized_critic.py for centralized critic PPO on this game. """ import argparse from gym.spaces import Tuple, MultiDiscrete,...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/__init__.py
Python
from ray.rllib.models.action_dist import ActionDistribution from ray.rllib.models.catalog import ModelCatalog, MODEL_DEFAULTS from ray.rllib.models.model import Model from ray.rllib.models.preprocessors import Preprocessor from ray.rllib.models.tf.fcnet_v1 import FullyConnectedNetwork from ray.rllib.models.tf.visionnet...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/action_dist.py
Python
from ray.rllib.utils.annotations import DeveloperAPI @DeveloperAPI class ActionDistribution: """The policy action distribution of an agent. Attributes: inputs (Tensors): input vector to compute samples from. model (ModelV2): reference to model producing the inputs. """ @DeveloperAPI ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/catalog.py
Python
import gym import logging import numpy as np from functools import partial from ray.tune.registry import RLLIB_MODEL, RLLIB_PREPROCESSOR, \ RLLIB_ACTION_DIST, _global_registry from ray.rllib.models.extra_spaces import Simplex from ray.rllib.models.torch.torch_action_dist import (TorchCategorical, ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/extra_spaces.py
Python
import numpy as np import gym class Simplex(gym.Space): """Represents a d - 1 dimensional Simplex in R^d. That is, all coordinates are in [0, 1] and sum to 1. The dimension d of the simplex is assumed to be shape[-1]. Additionally one can specify the underlying distribution of the simplex as a D...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/model.py
Python
from collections import OrderedDict import logging import gym from ray.rllib.models.tf.misc import linear, normc_initializer from ray.rllib.models.preprocessors import get_preprocessor from ray.rllib.utils.annotations import PublicAPI, DeveloperAPI from ray.rllib.utils import try_import_tf, try_import_torch tf = try_...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/modelv2.py
Python
from ray.rllib.policy.sample_batch import SampleBatch from ray.rllib.models.model import restore_original_dimensions, flatten from ray.rllib.utils.annotations import PublicAPI @PublicAPI class ModelV2: """Defines a Keras-style abstract network model for use with RLlib. Custom models should extend either TFMo...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/preprocessors.py
Python
from collections import OrderedDict import cv2 import logging import numpy as np import gym from ray.rllib.utils.annotations import override, PublicAPI ATARI_OBS_SHAPE = (210, 160, 3) ATARI_RAM_OBS_SHAPE = (128, ) VALIDATION_INTERVAL = 100 logger = logging.getLogger(__name__) @PublicAPI class Preprocessor: """...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/fcnet_v1.py
Python
from ray.rllib.models.model import Model from ray.rllib.models.tf.misc import normc_initializer, get_activation_fn from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_tf tf = try_import_tf() # Deprecated: see as an alternative models/tf/fcnet_v2.py class FullyConnectedNetwork(Mode...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/fcnet_v2.py
Python
import numpy as np from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.models.tf.misc import normc_initializer, get_activation_fn from ray.rllib.utils import try_import_tf tf = try_import_tf() class FullyConnectedNetwork(TFModelV2): """Generic fully connected network implemented in ModelV2 API."...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/lstm_v1.py
Python
import numpy as np from ray.rllib.models.model import Model from ray.rllib.models.tf.misc import linear, normc_initializer from ray.rllib.policy.rnn_sequencing import add_time_dimension from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_tf tf = try_import_tf() # Deprecated: see ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/misc.py
Python
import numpy as np from ray.rllib.utils import try_import_tf tf = try_import_tf() def normc_initializer(std=1.0): def _initializer(shape, dtype=None, partition_info=None): out = np.random.randn(*shape).astype(np.float32) out *= std / np.sqrt(np.square(out).sum(axis=0, keepdims=True)) retu...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/modelv1_compat.py
Python
import logging import numpy as np from ray.rllib.models.modelv2 import ModelV2 from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.models.tf.misc import linear, normc_initializer from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_tf from ray.rllib.utils.debug import...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/recurrent_tf_modelv2.py
Python
from ray.rllib.models.modelv2 import ModelV2 from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.policy.rnn_sequencing import add_time_dimension from ray.rllib.utils.annotations import override, DeveloperAPI from ray.rllib.utils import try_import_tf tf = try_import_tf() @DeveloperAPI class RecurrentT...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/tf_action_dist.py
Python
import numpy as np import functools from ray.rllib.models.action_dist import ActionDistribution from ray.rllib.policy.policy import TupleActions from ray.rllib.utils.annotations import override, DeveloperAPI from ray.rllib.utils import try_import_tf tf = try_import_tf() @DeveloperAPI class TFActionDistribution(Acti...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/tf_modelv2.py
Python
from ray.rllib.models.modelv2 import ModelV2 from ray.rllib.utils.annotations import PublicAPI from ray.rllib.utils import try_import_tf tf = try_import_tf() @PublicAPI class TFModelV2(ModelV2): """TF version of ModelV2. Note that this class by itself is not a valid model unless you implement forward() ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/visionnet_v1.py
Python
from ray.rllib.models.model import Model from ray.rllib.models.tf.misc import get_activation_fn, flatten from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_tf tf = try_import_tf() # Deprecated: see as an alternative models/tf/visionnet_v2.py class VisionNetwork(Model): """Gen...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/tf/visionnet_v2.py
Python
from ray.rllib.models.tf.tf_modelv2 import TFModelV2 from ray.rllib.models.tf.visionnet_v1 import _get_filter_config from ray.rllib.models.tf.misc import normc_initializer, get_activation_fn from ray.rllib.utils import try_import_tf tf = try_import_tf() class VisionNetwork(TFModelV2): """Generic vision network i...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/torch/fcnet.py
Python
import logging import numpy as np from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.models.torch.misc import normc_initializer, SlimFC, \ _get_activation_fn from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_torch _, nn = try_import_torch() logger =...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/torch/misc.py
Python
""" Code adapted from https://github.com/ikostrikov/pytorch-a3c""" import numpy as np from ray.rllib.utils import try_import_torch torch, nn = try_import_torch() def normc_initializer(std=1.0): def initializer(tensor): tensor.data.normal_(0, 1) tensor.data *= std / torch.sqrt( tensor...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/torch/torch_action_dist.py
Python
import numpy as np from ray.rllib.models.action_dist import ActionDistribution from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_torch torch, nn = try_import_torch() class TorchDistributionWrapper(ActionDistribution): """Wrapper class for torch.distributions.""" @overr...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/torch/torch_modelv2.py
Python
from ray.rllib.models.modelv2 import ModelV2 from ray.rllib.utils.annotations import PublicAPI from ray.rllib.utils import try_import_torch _, nn = try_import_torch() @PublicAPI class TorchModelV2(ModelV2): """Torch version of ModelV2. Note that this class by itself is not a valid model unless you inher...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/models/torch/visionnet.py
Python
from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.models.torch.misc import normc_initializer, valid_padding, \ SlimConv2d, SlimFC from ray.rllib.models.tf.visionnet_v1 import _get_filter_config from ray.rllib.utils.annotations import override from ray.rllib.utils import try_import_torch ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/offline/__init__.py
Python
from ray.rllib.offline.io_context import IOContext from ray.rllib.offline.json_reader import JsonReader from ray.rllib.offline.json_writer import JsonWriter from ray.rllib.offline.output_writer import OutputWriter, NoopOutput from ray.rllib.offline.input_reader import InputReader from ray.rllib.offline.mixed_input impo...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/offline/input_reader.py
Python
import logging import numpy as np import threading from ray.rllib.policy.sample_batch import MultiAgentBatch from ray.rllib.utils.annotations import PublicAPI from ray.rllib.utils import try_import_tf tf = try_import_tf() logger = logging.getLogger(__name__) @PublicAPI class InputReader: """Input object for lo...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/offline/io_context.py
Python
import os from ray.rllib.utils.annotations import PublicAPI @PublicAPI class IOContext: """Attributes to pass to input / output class constructors. RLlib auto-sets these attributes when constructing input / output classes. Attributes: log_dir (str): Default logging directory. config (di...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/offline/is_estimator.py
Python
from ray.rllib.offline.off_policy_estimator import OffPolicyEstimator, \ OffPolicyEstimate from ray.rllib.utils.annotations import override class ImportanceSamplingEstimator(OffPolicyEstimator): """The step-wise IS estimator. Step-wise IS estimator described in https://arxiv.org/pdf/1511.03722.pdf""" ...
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta
rllib/offline/json_reader.py
Python
import glob import json import logging import os import random import six from six.moves.urllib.parse import urlparse try: from smart_open import smart_open except ImportError: smart_open = None from ray.rllib.offline.input_reader import InputReader from ray.rllib.offline.io_context import IOContext from ray....
zhuohan123/hoplite-rllib
3
Python
zhuohan123
Zhuohan Li
vLLM / Meta