doc_content
stringlengths
1
386k
doc_id
stringlengths
5
188
property rpc_timeout A float indicating the timeout to use for all RPCs. If an RPC does not complete in this timeframe, it will complete with an exception indicating that it has timed out.
torch.rpc#torch.distributed.rpc.RpcBackendOptions.rpc_timeout
torch.distributed.rpc.rpc_async(to, func, args=None, kwargs=None, timeout=-1.0) [source] Make a non-blocking RPC call to run function func on worker to. RPC messages are sent and received in parallel to execution of Python code. This method is thread-safe. This method will immediately return a Future that can be awaited on. Parameters to (str or WorkerInfo or int) – name/rank/WorkerInfo of the destination worker. func (callable) – a callable function, such as Python callables, builtin operators (e.g. add()) and annotated TorchScript functions. args (tuple) – the argument tuple for the func invocation. kwargs (dict) – is a dictionary of keyword arguments for the func invocation. timeout (float, optional) – timeout in seconds to use for this RPC. If the RPC does not complete in this amount of time, an exception indicating it has timed out will be raised. A value of 0 indicates an infinite timeout, i.e. a timeout error will never be raised. If not provided, the default value set during initialization or with _set_rpc_timeout is used. Returns Returns a Future object that can be waited on. When completed, the return value of func on args and kwargs can be retrieved from the Future object. Warning Using GPU tensors as arguments or return values of func is not supported since we don’t support sending GPU tensors over the wire. You need to explicitly copy GPU tensors to CPU before using them as arguments or return values of func. Warning The rpc_async API does not copy storages of argument tensors until sending them over the wire, which could be done by a different thread depending on the RPC backend type. The caller should make sure that the contents of those tensors stay intact until the returned Future completes. Example:: Make sure that MASTER_ADDR and MASTER_PORT are set properly on both workers. Refer to init_process_group() API for more details. For example, >>> export MASTER_ADDR=localhost >>> export MASTER_PORT=5678 Then run the following code in two different processes: >>> # On worker 0: >>> import torch >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker0", rank=0, world_size=2) >>> fut1 = rpc.rpc_async("worker1", torch.add, args=(torch.ones(2), 3)) >>> fut2 = rpc.rpc_async("worker1", min, args=(1, 2)) >>> result = fut1.wait() + fut2.wait() >>> rpc.shutdown() >>> # On worker 1: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker1", rank=1, world_size=2) >>> rpc.shutdown() Below is an example of running a TorchScript function using RPC. >>> # On both workers: >>> @torch.jit.script >>> def my_script_add(t1, t2): >>> return torch.add(t1, t2) >>> # On worker 0: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker0", rank=0, world_size=2) >>> fut = rpc.rpc_async("worker1", my_script_add, args=(torch.ones(2), 3)) >>> ret = fut.wait() >>> rpc.shutdown() >>> # On worker 1: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker1", rank=1, world_size=2) >>> rpc.shutdown()
torch.rpc#torch.distributed.rpc.rpc_async
torch.distributed.rpc.rpc_sync(to, func, args=None, kwargs=None, timeout=-1.0) [source] Make a blocking RPC call to run function func on worker to. RPC messages are sent and received in parallel to execution of Python code. This method is thread-safe. Parameters to (str or WorkerInfo or int) – name/rank/WorkerInfo of the destination worker. func (callable) – a callable function, such as Python callables, builtin operators (e.g. add()) and annotated TorchScript functions. args (tuple) – the argument tuple for the func invocation. kwargs (dict) – is a dictionary of keyword arguments for the func invocation. timeout (float, optional) – timeout in seconds to use for this RPC. If the RPC does not complete in this amount of time, an exception indicating it has timed out will be raised. A value of 0 indicates an infinite timeout, i.e. a timeout error will never be raised. If not provided, the default value set during initialization or with _set_rpc_timeout is used. Returns Returns the result of running func with args and kwargs. Warning Using GPU tensors as arguments or return values of func is not supported since we don’t support sending GPU tensors over the wire. You need to explicitly copy GPU tensors to CPU before using them as arguments or return values of func. Example:: Make sure that MASTER_ADDR and MASTER_PORT are set properly on both workers. Refer to init_process_group() API for more details. For example, >>> export MASTER_ADDR=localhost >>> export MASTER_PORT=5678 Then run the following code in two different processes: >>> # On worker 0: >>> import torch >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker0", rank=0, world_size=2) >>> ret = rpc.rpc_sync("worker1", torch.add, args=(torch.ones(2), 3)) >>> rpc.shutdown() >>> # On worker 1: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker1", rank=1, world_size=2) >>> rpc.shutdown() Below is an example of running a TorchScript function using RPC. >>> # On both workers: >>> @torch.jit.script >>> def my_script_add(t1, t2): >>> return torch.add(t1, t2) >>> # On worker 0: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker0", rank=0, world_size=2) >>> ret = rpc.rpc_sync("worker1", my_script_add, args=(torch.ones(2), 3)) >>> rpc.shutdown() >>> # On worker 1: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker1", rank=1, world_size=2) >>> rpc.shutdown()
torch.rpc#torch.distributed.rpc.rpc_sync
class torch.distributed.rpc.RRef [source] backward(self: torch._C._distributed_rpc.PyRRef, dist_autograd_ctx_id: int = -1, retain_graph: bool = False) → None Runs the backward pass using the RRef as the root of the backward pass. If dist_autograd_ctx_id is provided, we perform a distributed backward pass using the provided ctx_id starting from the owner of the RRef. In this case, get_gradients() should be used to retrieve the gradients. If dist_autograd_ctx_id is None, it is assumed that this is a local autograd graph and we only perform a local backward pass. In the local case, the node calling this API has to be the owner of the RRef. The value of the RRef is expected to be a scalar Tensor. Parameters dist_autograd_ctx_id (int, optional) – The distributed autograd context id for which we should retrieve the gradients (default: -1). retain_graph (bool, optional) – If False, the graph used to compute the grad will be freed. Note that in nearly all cases setting this option to True is not needed and often can be worked around in a much more efficient way. Usually, you need to set this to True to run backward multiple times (default: False). Example:: >>> import torch.distributed.autograd as dist_autograd >>> with dist_autograd.context() as context_id: >>> rref.backward(context_id) confirmed_by_owner(self: torch._C._distributed_rpc.PyRRef) → bool Returns whether this RRef has been confirmed by the owner. OwnerRRef always returns true, while UserRRef only returns true when the owner knowns about this UserRRef. is_owner(self: torch._C._distributed_rpc.PyRRef) → bool Returns whether or not the current node is the owner of this RRef. local_value(self: torch._C._distributed_rpc.PyRRef) → object If the current node is the owner, returns a reference to the local value. Otherwise, throws an exception. owner(self: torch._C._distributed_rpc.PyRRef) → torch._C._distributed_rpc.WorkerInfo Returns worker information of the node that owns this RRef. owner_name(self: torch._C._distributed_rpc.PyRRef) → str Returns worker name of the node that owns this RRef. remote(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch a remote using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.remote().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.remote(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.remote(). If the creation of this RRef is not successfully completed within the timeout, then the next time there is an attempt to use the RRef (such as to_here), a timeout will be raised. If not provided, the default RPC timeout will be used. Please see rpc.remote() for specific timeout semantics for RRef. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.remote().size().to_here() # returns torch.Size([2, 2]) >>> rref.remote().view(1, 4).to_here() # returns tensor([[1., 1., 1., 1.]]) rpc_async(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch an rpc_async using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.rpc_async().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.rpc_async(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.rpc_async(). If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout will be used. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.rpc_async().size().wait() # returns torch.Size([2, 2]) >>> rref.rpc_async().view(1, 4).wait() # returns tensor([[1., 1., 1., 1.]]) rpc_sync(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch an rpc_sync using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.rpc_sync().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.rpc_sync(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.rpc_sync(). If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout will be used. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.rpc_sync().size() # returns torch.Size([2, 2]) >>> rref.rpc_sync().view(1, 4) # returns tensor([[1., 1., 1., 1.]]) to_here(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Blocking call that copies the value of the RRef from the owner to the local node and returns it. If the current node is the owner, returns a reference to the local value. Parameters timeout (float, optional) – Timeout for to_here. If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout (60s) will be used.
torch.rpc#torch.distributed.rpc.RRef
backward(self: torch._C._distributed_rpc.PyRRef, dist_autograd_ctx_id: int = -1, retain_graph: bool = False) → None Runs the backward pass using the RRef as the root of the backward pass. If dist_autograd_ctx_id is provided, we perform a distributed backward pass using the provided ctx_id starting from the owner of the RRef. In this case, get_gradients() should be used to retrieve the gradients. If dist_autograd_ctx_id is None, it is assumed that this is a local autograd graph and we only perform a local backward pass. In the local case, the node calling this API has to be the owner of the RRef. The value of the RRef is expected to be a scalar Tensor. Parameters dist_autograd_ctx_id (int, optional) – The distributed autograd context id for which we should retrieve the gradients (default: -1). retain_graph (bool, optional) – If False, the graph used to compute the grad will be freed. Note that in nearly all cases setting this option to True is not needed and often can be worked around in a much more efficient way. Usually, you need to set this to True to run backward multiple times (default: False). Example:: >>> import torch.distributed.autograd as dist_autograd >>> with dist_autograd.context() as context_id: >>> rref.backward(context_id)
torch.rpc#torch.distributed.rpc.RRef.backward
confirmed_by_owner(self: torch._C._distributed_rpc.PyRRef) → bool Returns whether this RRef has been confirmed by the owner. OwnerRRef always returns true, while UserRRef only returns true when the owner knowns about this UserRRef.
torch.rpc#torch.distributed.rpc.RRef.confirmed_by_owner
is_owner(self: torch._C._distributed_rpc.PyRRef) → bool Returns whether or not the current node is the owner of this RRef.
torch.rpc#torch.distributed.rpc.RRef.is_owner
local_value(self: torch._C._distributed_rpc.PyRRef) → object If the current node is the owner, returns a reference to the local value. Otherwise, throws an exception.
torch.rpc#torch.distributed.rpc.RRef.local_value
owner(self: torch._C._distributed_rpc.PyRRef) → torch._C._distributed_rpc.WorkerInfo Returns worker information of the node that owns this RRef.
torch.rpc#torch.distributed.rpc.RRef.owner
owner_name(self: torch._C._distributed_rpc.PyRRef) → str Returns worker name of the node that owns this RRef.
torch.rpc#torch.distributed.rpc.RRef.owner_name
remote(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch a remote using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.remote().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.remote(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.remote(). If the creation of this RRef is not successfully completed within the timeout, then the next time there is an attempt to use the RRef (such as to_here), a timeout will be raised. If not provided, the default RPC timeout will be used. Please see rpc.remote() for specific timeout semantics for RRef. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.remote().size().to_here() # returns torch.Size([2, 2]) >>> rref.remote().view(1, 4).to_here() # returns tensor([[1., 1., 1., 1.]])
torch.rpc#torch.distributed.rpc.RRef.remote
rpc_async(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch an rpc_async using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.rpc_async().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.rpc_async(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.rpc_async(). If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout will be used. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.rpc_async().size().wait() # returns torch.Size([2, 2]) >>> rref.rpc_async().view(1, 4).wait() # returns tensor([[1., 1., 1., 1.]])
torch.rpc#torch.distributed.rpc.RRef.rpc_async
rpc_sync(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Create a helper proxy to easily launch an rpc_sync using the owner of the RRef as the destination to run functions on the object referenced by this RRef. More specifically, rref.rpc_sync().func_name(*args, **kwargs) is the same as the following: >>> def run(rref, func_name, args, kwargs): >>> return getattr(rref.local_value(), func_name)(*args, **kwargs) >>> >>> rpc.rpc_sync(rref.owner(), run, args=(rref, func_name, args, kwargs)) Parameters timeout (float, optional) – Timeout for rref.rpc_sync(). If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout will be used. Example:: >>> from torch.distributed import rpc >>> rref = rpc.remote("worker1", torch.add, args=(torch.zeros(2, 2), 1)) >>> rref.rpc_sync().size() # returns torch.Size([2, 2]) >>> rref.rpc_sync().view(1, 4) # returns tensor([[1., 1., 1., 1.]])
torch.rpc#torch.distributed.rpc.RRef.rpc_sync
to_here(self: torch._C._distributed_rpc.PyRRef, timeout: float = -1.0) → object Blocking call that copies the value of the RRef from the owner to the local node and returns it. If the current node is the owner, returns a reference to the local value. Parameters timeout (float, optional) – Timeout for to_here. If the call does not complete within this timeframe, an exception indicating so will be raised. If this argument is not provided, the default RPC timeout (60s) will be used.
torch.rpc#torch.distributed.rpc.RRef.to_here
torch.distributed.rpc.shutdown(graceful=True) [source] Perform a shutdown of the RPC agent, and then destroy the RPC agent. This stops the local agent from accepting outstanding requests, and shuts down the RPC framework by terminating all RPC threads. If graceful=True, this will block until all local and remote RPC processes reach this method and wait for all outstanding work to complete. Otherwise, if graceful=False, this is a local shutdown, and it does not wait for other RPC processes to reach this method. Warning For Future objects returned by rpc_async(), future.wait() should not be called after shutdown(). Parameters graceful (bool) – Whether to do a graceful shutdown or not. If True, this will 1) wait until there is no pending system messages for UserRRefs and delete them; 2) block until all local and remote RPC processes have reached this method and wait for all outstanding work to complete. Example:: Make sure that MASTER_ADDR and MASTER_PORT are set properly on both workers. Refer to init_process_group() API for more details. For example, >>> export MASTER_ADDR=localhost >>> export MASTER_PORT=5678 Then run the following code in two different processes: >>> # On worker 0: >>> import torch >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker0", rank=0, world_size=2) >>> # do some work >>> result = rpc.rpc_sync("worker1", torch.add, args=(torch.ones(1), 1)) >>> # ready to shutdown >>> rpc.shutdown() >>> # On worker 1: >>> import torch.distributed.rpc as rpc >>> rpc.init_rpc("worker1", rank=1, world_size=2) >>> # wait for worker 0 to finish work, and then shutdown. >>> rpc.shutdown()
torch.rpc#torch.distributed.rpc.shutdown
class torch.distributed.rpc.TensorPipeRpcBackendOptions(*, num_worker_threads=16, rpc_timeout=60.0, init_method='env://', device_maps=None, _transports=None, _channels=None) [source] The backend options for TensorPipeAgent, derived from RpcBackendOptions. Parameters num_worker_threads (int, optional) – The number of threads in the thread-pool used by TensorPipeAgent to execute requests (default: 16). rpc_timeout (float, optional) – The default timeout, in seconds, for RPC requests (default: 60 seconds). If the RPC has not completed in this timeframe, an exception indicating so will be raised. Callers can override this timeout for individual RPCs in rpc_sync() and rpc_async() if necessary. init_method (str, optional) – The URL to initialize the distributed store used for rendezvous. It takes any value accepted for the same argument of init_process_group() (default: env://). device_maps (Dict[str, Dict]) – Device placement mappings from this worker to the callee. Key is the callee worker name and value the dictionary (Dict of int, str, or torch.device) that maps this worker’s devices to the callee worker’s devices. (default: None) property device_maps The device map locations. property init_method URL specifying how to initialize the process group. Default is env:// property num_worker_threads The number of threads in the thread-pool used by TensorPipeAgent to execute requests. property rpc_timeout A float indicating the timeout to use for all RPCs. If an RPC does not complete in this timeframe, it will complete with an exception indicating that it has timed out. set_device_map(to, device_map) [source] Set device mapping between each RPC caller and callee pair. This function can be called multiple times to incrementally add device placement configurations. Parameters worker_name (str) – Callee name. device_map (Dict of python:int, str, or torch.device) – Device placement mappings from this worker to the callee. This map must be invertible. Example:: >>> # both workers >>> def add(x, y): >>> print(x) # tensor([1., 1.], device='cuda:1') >>> return x + y, (x + y).to(2) >>> >>> # on worker 0 >>> options = TensorPipeRpcBackendOptions( >>> num_worker_threads=8, >>> device_maps={"worker1": {0, 1}} >>> # maps worker0's cuda:0 to worker1's cuda:1 >>> ) >>> options.set_device_map("worker1", {1, 2}) >>> # maps worker0's cuda:1 to worker1's cuda:2 >>> >>> rpc.init_rpc( >>> "worker0", >>> rank=0, >>> world_size=2 >>> backend=rpc.BackendType.TENSORPIPE, >>> rpc_backend_options=options >>> ) >>> >>> x = torch.ones(2) >>> rets = rpc.rpc_sync("worker1", add, args=(x.to(0), 1)) >>> # The first argument will be moved to cuda:1 on worker1. When >>> # sending the return value back, it will follow the invert of >>> # the device map, and hence will be moved back to cuda:0 and >>> # cuda:1 on worker0 >>> print(rets[0]) # tensor([2., 2.], device='cuda:0') >>> print(rets[0]) # tensor([2., 2.], device='cuda:1')
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions
property device_maps The device map locations.
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions.device_maps
property init_method URL specifying how to initialize the process group. Default is env://
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions.init_method
property num_worker_threads The number of threads in the thread-pool used by TensorPipeAgent to execute requests.
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions.num_worker_threads
property rpc_timeout A float indicating the timeout to use for all RPCs. If an RPC does not complete in this timeframe, it will complete with an exception indicating that it has timed out.
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions.rpc_timeout
set_device_map(to, device_map) [source] Set device mapping between each RPC caller and callee pair. This function can be called multiple times to incrementally add device placement configurations. Parameters worker_name (str) – Callee name. device_map (Dict of python:int, str, or torch.device) – Device placement mappings from this worker to the callee. This map must be invertible. Example:: >>> # both workers >>> def add(x, y): >>> print(x) # tensor([1., 1.], device='cuda:1') >>> return x + y, (x + y).to(2) >>> >>> # on worker 0 >>> options = TensorPipeRpcBackendOptions( >>> num_worker_threads=8, >>> device_maps={"worker1": {0, 1}} >>> # maps worker0's cuda:0 to worker1's cuda:1 >>> ) >>> options.set_device_map("worker1", {1, 2}) >>> # maps worker0's cuda:1 to worker1's cuda:2 >>> >>> rpc.init_rpc( >>> "worker0", >>> rank=0, >>> world_size=2 >>> backend=rpc.BackendType.TENSORPIPE, >>> rpc_backend_options=options >>> ) >>> >>> x = torch.ones(2) >>> rets = rpc.rpc_sync("worker1", add, args=(x.to(0), 1)) >>> # The first argument will be moved to cuda:1 on worker1. When >>> # sending the return value back, it will follow the invert of >>> # the device map, and hence will be moved back to cuda:0 and >>> # cuda:1 on worker0 >>> print(rets[0]) # tensor([2., 2.], device='cuda:0') >>> print(rets[0]) # tensor([2., 2.], device='cuda:1')
torch.rpc#torch.distributed.rpc.TensorPipeRpcBackendOptions.set_device_map
class torch.distributed.rpc.WorkerInfo A structure that encapsulates information of a worker in the system. Contains the name and ID of the worker. This class is not meant to be constructed directly, rather, an instance can be retrieved through get_worker_info() and the result can be passed in to functions such as rpc_sync(), rpc_async(), remote() to avoid copying a string on every invocation. property id Globally unique id to identify the worker. property name The name of the worker.
torch.rpc#torch.distributed.rpc.WorkerInfo
property id Globally unique id to identify the worker.
torch.rpc#torch.distributed.rpc.WorkerInfo.id
property name The name of the worker.
torch.rpc#torch.distributed.rpc.WorkerInfo.name
torch.distributed.scatter(tensor, scatter_list=None, src=0, group=None, async_op=False) [source] Scatters a list of tensors to all processes in a group. Each process will receive exactly one tensor and store its data in the tensor argument. Parameters tensor (Tensor) – Output tensor. scatter_list (list[Tensor]) – List of tensors to scatter (default is None, must be specified on the source rank) src (int) – Source rank (default is 0) group (ProcessGroup, optional) – The process group to work on. If None, the default process group will be used. async_op (bool, optional) – Whether this op should be an async op Returns Async work handle, if async_op is set to True. None, if not async_op or if not part of the group
torch.distributed#torch.distributed.scatter
torch.distributed.scatter_object_list(scatter_object_output_list, scatter_object_input_list, src=0, group=None) [source] Scatters picklable objects in scatter_object_input_list to the whole group. Similar to scatter(), but Python objects can be passed in. On each rank, the scattered object will be stored as the first element of scatter_object_output_list. Note that all objects in scatter_object_input_list must be picklable in order to be scattered. Parameters scatter_object_output_list (List[Any]) – Non-empty list whose first element will store the object scattered to this rank. scatter_object_input_list (List[Any]) – List of input objects to scatter. Each object must be picklable. Only objects on the src rank will be scattered, and the argument can be None for non-src ranks. src (int) – Source rank from which to scatter scatter_object_input_list. group – (ProcessGroup, optional): The process group to work on. If None, the default process group will be used. Default is None. Returns None. If rank is part of the group, scatter_object_output_list will have its first element set to the scattered object for this rank. Note Note that this API differs slightly from the scatter collective since it does not provide an async_op handle and thus will be a blocking call. Warning scatter_object_list() uses pickle module implicitly, which is known to be insecure. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling. Only call this function with data you trust. Example:: >>> # Note: Process group initialization omitted on each rank. >>> import torch.distributed as dist >>> if dist.get_rank() == 0: >>> # Assumes world_size of 3. >>> objects = ["foo", 12, {1: 2}] # any picklable object >>> else: >>> # Can be any list on non-src ranks, elements are not used. >>> objects = [None, None, None] >>> output_list = [None] >>> dist.scatter_object_list(output_list, objects, src=0) >>> # Rank i gets objects[i]. For example, on rank 2: >>> output_list [{1: 2}]
torch.distributed#torch.distributed.scatter_object_list
torch.distributed.send(tensor, dst, group=None, tag=0) [source] Sends a tensor synchronously. Parameters tensor (Tensor) – Tensor to send. dst (int) – Destination rank. group (ProcessGroup, optional) – The process group to work on. If None, the default process group will be used. tag (int, optional) – Tag to match send with remote recv
torch.distributed#torch.distributed.send
class torch.distributed.Store Base class for all store implementations, such as the 3 provided by PyTorch distributed: (TCPStore, FileStore, and HashStore).
torch.distributed#torch.distributed.Store
torch.distributed.Store.add(self: torch._C._distributed_c10d.Store, arg0: str, arg1: int) → int The first call to add for a given key creates a counter associated with key in the store, initialized to amount. Subsequent calls to add with the same key increment the counter by the specified amount. Calling add() with a key that has already been set in the store by set() will result in an exception. Parameters key (str) – The key in the store whose counter will be incremented. amount (int) – The quantity by which the counter will be incremented. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, other store types can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.add("first_key", 1) >>> store.add("first_key", 6) >>> # Should return 7 >>> store.get("first_key")
torch.distributed#torch.distributed.Store.add
torch.distributed.Store.delete_key(self: torch._C._distributed_c10d.Store, arg0: str) → bool Deletes the key-value pair associated with key from the store. Returns true if the key was successfully deleted, and false if it was not. Warning The delete_key API is only supported by the TCPStore and HashStore. Using this API with the FileStore will result in an exception. Parameters key (str) – The key to be deleted from the store Returns True if key was deleted, otherwise False. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, HashStore can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.set("first_key") >>> # This should return true >>> store.delete_key("first_key") >>> # This should return false >>> store.delete_key("bad_key")
torch.distributed#torch.distributed.Store.delete_key
torch.distributed.Store.get(self: torch._C._distributed_c10d.Store, arg0: str) → bytes Retrieves the value associated with the given key in the store. If key is not present in the store, the function will wait for timeout, which is defined when initializing the store, before throwing an exception. Parameters key (str) – The function will return the value associated with this key. Returns Value associated with key if key is in the store. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.set("first_key", "first_value") >>> # Should return "first_value" >>> store.get("first_key")
torch.distributed#torch.distributed.Store.get
torch.distributed.Store.num_keys(self: torch._C._distributed_c10d.Store) → int Returns the number of keys set in the store. Note that this number will typically be one greater than the number of keys added by set() and add() since one key is used to coordinate all the workers using the store. Warning When used with the TCPStore, num_keys returns the number of keys written to the underlying file. If the store is destructed and another store is created with the same file, the original keys will be retained. Returns The number of keys present in the store. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, other store types can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.set("first_key", "first_value") >>> # This should return 2 >>> store.num_keys()
torch.distributed#torch.distributed.Store.num_keys
torch.distributed.Store.set(self: torch._C._distributed_c10d.Store, arg0: str, arg1: str) → None Inserts the key-value pair into the store based on the supplied key and value. If key already exists in the store, it will overwrite the old value with the new supplied value. Parameters key (str) – The key to be added to the store. value (str) – The value associated with key to be added to the store. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.set("first_key", "first_value") >>> # Should return "first_value" >>> store.get("first_key")
torch.distributed#torch.distributed.Store.set
torch.distributed.Store.set_timeout(self: torch._C._distributed_c10d.Store, arg0: datetime.timedelta) → None Sets the store’s default timeout. This timeout is used during initialization and in wait() and get(). Parameters timeout (timedelta) – timeout to be set in the store. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, other store types can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> store.set_timeout(timedelta(seconds=10)) >>> # This will throw an exception after 10 seconds >>> store.wait(["bad_key"])
torch.distributed#torch.distributed.Store.set_timeout
torch.distributed.Store.wait(*args, **kwargs) Overloaded function. wait(self: torch._C._distributed_c10d.Store, arg0: List[str]) -> None Waits for each key in keys to be added to the store. If not all keys are set before the timeout (set during store initialization), then wait will throw an exception. Parameters keys (list) – List of keys on which to wait until they are set in the store. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, other store types can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> # This will throw an exception after 30 seconds >>> store.wait(["bad_key"]) wait(self: torch._C._distributed_c10d.Store, arg0: List[str], arg1: datetime.timedelta) -> None Waits for each key in keys to be added to the store, and throws an exception if the keys have not been set by the supplied timeout. Parameters keys (list) – List of keys on which to wait until they are set in the store. timeout (timedelta) – Time to wait for the keys to be added before throwing an exception. Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Using TCPStore as an example, other store types can also be used >>> store = dist.TCPStore("127.0.0.1", 0, 1, True, timedelta(seconds=30)) >>> # This will throw an exception after 10 seconds >>> store.wait(["bad_key"], timedelta(seconds=10))
torch.distributed#torch.distributed.Store.wait
class torch.distributed.TCPStore A TCP-based distributed key-value store implementation. The server store holds the data, while the client stores can connect to the server store over TCP and perform actions such as set() to insert a key-value pair, get() to retrieve a key-value pair, etc. Parameters host_name (str) – The hostname or IP Address the server store should run on. port (int) – The port on which the server store should listen for incoming requests. world_size (int) – The total number of store users (number of clients + 1 for the server). is_master (bool) – True when initializing the server store, False for client stores. timeout (timedelta) – Timeout used by the store during initialization and for methods such as get() and wait(). Example:: >>> import torch.distributed as dist >>> from datetime import timedelta >>> # Run on process 1 (server) >>> server_store = dist.TCPStore("127.0.0.1", 1234, 2, True, timedelta(seconds=30)) >>> # Run on process 2 (client) >>> client_store = dist.TCPStore("127.0.0.1", 1234, 2, False) >>> # Use any of the store methods from either the client or server after initialization >>> server_store.set("first_key", "first_value") >>> client_store.get("first_key")
torch.distributed#torch.distributed.TCPStore
Probability distributions - torch.distributions The distributions package contains parameterizable probability distributions and sampling functions. This allows the construction of stochastic computation graphs and stochastic gradient estimators for optimization. This package generally follows the design of the TensorFlow Distributions package. It is not possible to directly backpropagate through random samples. However, there are two main methods for creating surrogate functions that can be backpropagated through. These are the score function estimator/likelihood ratio estimator/REINFORCE and the pathwise derivative estimator. REINFORCE is commonly seen as the basis for policy gradient methods in reinforcement learning, and the pathwise derivative estimator is commonly seen in the reparameterization trick in variational autoencoders. Whilst the score function only requires the value of samples f(x)f(x) , the pathwise derivative requires the derivative f′(x)f'(x) . The next sections discuss these two in a reinforcement learning example. For more details see Gradient Estimation Using Stochastic Computation Graphs . Score function When the probability density function is differentiable with respect to its parameters, we only need sample() and log_prob() to implement REINFORCE: Δθ=αr∂log⁡p(a∣πθ(s))∂θ\Delta\theta = \alpha r \frac{\partial\log p(a|\pi^\theta(s))}{\partial\theta} where θ\theta are the parameters, α\alpha is the learning rate, rr is the reward and p(a∣πθ(s))p(a|\pi^\theta(s)) is the probability of taking action aa in state ss given policy πθ\pi^\theta . In practice we would sample an action from the output of a network, apply this action in an environment, and then use log_prob to construct an equivalent loss function. Note that we use a negative because optimizers use gradient descent, whilst the rule above assumes gradient ascent. With a categorical policy, the code for implementing REINFORCE would be as follows: probs = policy_network(state) # Note that this is equivalent to what used to be called multinomial m = Categorical(probs) action = m.sample() next_state, reward = env.step(action) loss = -m.log_prob(action) * reward loss.backward() Pathwise derivative The other way to implement these stochastic/policy gradients would be to use the reparameterization trick from the rsample() method, where the parameterized random variable can be constructed via a parameterized deterministic function of a parameter-free random variable. The reparameterized sample therefore becomes differentiable. The code for implementing the pathwise derivative would be as follows: params = policy_network(state) m = Normal(*params) # Any distribution with .has_rsample == True could work based on the application action = m.rsample() next_state, reward = env.step(action) # Assuming that reward is differentiable loss = -reward loss.backward() Distribution class torch.distributions.distribution.Distribution(batch_shape=torch.Size([]), event_shape=torch.Size([]), validate_args=None) [source] Bases: object Distribution is the abstract base class for probability distributions. property arg_constraints Returns a dictionary from argument names to Constraint objects that should be satisfied by each argument of this distribution. Args that are not tensors need not appear in this dict. property batch_shape Returns the shape over which parameters are batched. cdf(value) [source] Returns the cumulative density/mass function evaluated at value. Parameters value (Tensor) – entropy() [source] Returns entropy of distribution, batched over batch_shape. Returns Tensor of shape batch_shape. enumerate_support(expand=True) [source] Returns tensor containing all values supported by a discrete distribution. The result will enumerate over dimension 0, so the shape of the result will be (cardinality,) + batch_shape + event_shape (where event_shape = () for univariate distributions). Note that this enumerates over all batched tensors in lock-step [[0, 0], [1, 1], …]. With expand=False, enumeration happens along dim 0, but with the remaining batch dimensions being singleton dimensions, [[0], [1], ... To iterate over the full Cartesian product use itertools.product(m.enumerate_support()). Parameters expand (bool) – whether to expand the support over the batch dims to match the distribution’s batch_shape. Returns Tensor iterating over dimension 0. property event_shape Returns the shape of a single sample (without batching). expand(batch_shape, _instance=None) [source] Returns a new distribution instance (or populates an existing instance provided by a derived class) with batch dimensions expanded to batch_shape. This method calls expand on the distribution’s parameters. As such, this does not allocate new memory for the expanded distribution instance. Additionally, this does not repeat any args checking or parameter broadcasting in __init__.py, when an instance is first created. Parameters batch_shape (torch.Size) – the desired expanded size. _instance – new instance provided by subclasses that need to override .expand. Returns New distribution instance with batch dimensions expanded to batch_size. icdf(value) [source] Returns the inverse cumulative density/mass function evaluated at value. Parameters value (Tensor) – log_prob(value) [source] Returns the log of the probability density/mass function evaluated at value. Parameters value (Tensor) – property mean Returns the mean of the distribution. perplexity() [source] Returns perplexity of distribution, batched over batch_shape. Returns Tensor of shape batch_shape. rsample(sample_shape=torch.Size([])) [source] Generates a sample_shape shaped reparameterized sample or sample_shape shaped batch of reparameterized samples if the distribution parameters are batched. sample(sample_shape=torch.Size([])) [source] Generates a sample_shape shaped sample or sample_shape shaped batch of samples if the distribution parameters are batched. sample_n(n) [source] Generates n samples or n batches of samples if the distribution parameters are batched. static set_default_validate_args(value) [source] Sets whether validation is enabled or disabled. The default behavior mimics Python’s assert statement: validation is on by default, but is disabled if Python is run in optimized mode (via python -O). Validation may be expensive, so you may want to disable it once a model is working. Parameters value (bool) – Whether to enable validation. property stddev Returns the standard deviation of the distribution. property support Returns a Constraint object representing this distribution’s support. property variance Returns the variance of the distribution. ExponentialFamily class torch.distributions.exp_family.ExponentialFamily(batch_shape=torch.Size([]), event_shape=torch.Size([]), validate_args=None) [source] Bases: torch.distributions.distribution.Distribution ExponentialFamily is the abstract base class for probability distributions belonging to an exponential family, whose probability mass/density function has the form is defined below pF(x;θ)=exp⁡(⟨t(x),θ⟩−F(θ)+k(x))p_{F}(x; \theta) = \exp(\langle t(x), \theta\rangle - F(\theta) + k(x)) where θ\theta denotes the natural parameters, t(x)t(x) denotes the sufficient statistic, F(θ)F(\theta) is the log normalizer function for a given family and k(x)k(x) is the carrier measure. Note This class is an intermediary between the Distribution class and distributions which belong to an exponential family mainly to check the correctness of the .entropy() and analytic KL divergence methods. We use this class to compute the entropy and KL divergence using the AD framework and Bregman divergences (courtesy of: Frank Nielsen and Richard Nock, Entropies and Cross-entropies of Exponential Families). entropy() [source] Method to compute the entropy using Bregman divergence of the log normalizer. Bernoulli class torch.distributions.bernoulli.Bernoulli(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Bernoulli distribution parameterized by probs or logits (but not both). Samples are binary (0 or 1). They take the value 1 with probability p and 0 with probability 1 - p. Example: >>> m = Bernoulli(torch.tensor([0.3])) >>> m.sample() # 30% chance 1; 70% chance 0 tensor([ 0.]) Parameters probs (Number, Tensor) – the probability of sampling 1 logits (Number, Tensor) – the log-odds of sampling 1 arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] support = Boolean() property variance Beta class torch.distributions.beta.Beta(concentration1, concentration0, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Beta distribution parameterized by concentration1 and concentration0. Example: >>> m = Beta(torch.tensor([0.5]), torch.tensor([0.5])) >>> m.sample() # Beta distributed with concentration concentration1 and concentration0 tensor([ 0.1046]) Parameters concentration1 (float or Tensor) – 1st concentration parameter of the distribution (often referred to as alpha) concentration0 (float or Tensor) – 2nd concentration parameter of the distribution (often referred to as beta) arg_constraints = {'concentration0': GreaterThan(lower_bound=0.0), 'concentration1': GreaterThan(lower_bound=0.0)} property concentration0 property concentration1 entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=()) [source] support = Interval(lower_bound=0.0, upper_bound=1.0) property variance Binomial class torch.distributions.binomial.Binomial(total_count=1, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Binomial distribution parameterized by total_count and either probs or logits (but not both). total_count must be broadcastable with probs/logits. Example: >>> m = Binomial(100, torch.tensor([0 , .2, .8, 1])) >>> x = m.sample() tensor([ 0., 22., 71., 100.]) >>> m = Binomial(torch.tensor([[5.], [10.]]), torch.tensor([0.5, 0.8])) >>> x = m.sample() tensor([[ 4., 5.], [ 7., 6.]]) Parameters total_count (int or Tensor) – number of Bernoulli trials probs (Tensor) – Event probabilities logits (Tensor) – Event log-odds arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0), 'total_count': IntegerGreaterThan(lower_bound=0)} enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] property support property variance Categorical class torch.distributions.categorical.Categorical(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a categorical distribution parameterized by either probs or logits (but not both). Note It is equivalent to the distribution that torch.multinomial() samples from. Samples are integers from {0,…,K−1}\{0, \ldots, K-1\} where K is probs.size(-1). If probs is 1-dimensional with length-K, each element is the relative probability of sampling the class at that index. If probs is N-dimensional, the first N-1 dimensions are treated as a batch of relative probability vectors. Note The probs argument must be non-negative, finite and have a non-zero sum, and it will be normalized to sum to 1 along the last dimension. attr:probs will return this normalized value. The logits argument will be interpreted as unnormalized log probabilities and can therefore be any real number. It will likewise be normalized so that the resulting probabilities sum to 1 along the last dimension. attr:logits will return this normalized value. See also: torch.multinomial() Example: >>> m = Categorical(torch.tensor([ 0.25, 0.25, 0.25, 0.25 ])) >>> m.sample() # equal probability of 0, 1, 2, 3 tensor(3) Parameters probs (Tensor) – event probabilities logits (Tensor) – event log probabilities (unnormalized) arg_constraints = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] property support property variance Cauchy class torch.distributions.cauchy.Cauchy(loc, scale, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Samples from a Cauchy (Lorentz) distribution. The distribution of the ratio of independent normally distributed random variables with means 0 follows a Cauchy distribution. Example: >>> m = Cauchy(torch.tensor([0.0]), torch.tensor([1.0])) >>> m.sample() # sample from a Cauchy distribution with loc=0 and scale=1 tensor([ 2.3214]) Parameters loc (float or Tensor) – mode or median of the distribution. scale (float or Tensor) – half width at half maximum. arg_constraints = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] support = Real() property variance Chi2 class torch.distributions.chi2.Chi2(df, validate_args=None) [source] Bases: torch.distributions.gamma.Gamma Creates a Chi2 distribution parameterized by shape parameter df. This is exactly equivalent to Gamma(alpha=0.5*df, beta=0.5) Example: >>> m = Chi2(torch.tensor([1.0])) >>> m.sample() # Chi2 distributed with shape df=1 tensor([ 0.1046]) Parameters df (float or Tensor) – shape parameter of the distribution arg_constraints = {'df': GreaterThan(lower_bound=0.0)} property df expand(batch_shape, _instance=None) [source] ContinuousBernoulli class torch.distributions.continuous_bernoulli.ContinuousBernoulli(probs=None, logits=None, lims=(0.499, 0.501), validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a continuous Bernoulli distribution parameterized by probs or logits (but not both). The distribution is supported in [0, 1] and parameterized by ‘probs’ (in (0,1)) or ‘logits’ (real-valued). Note that, unlike the Bernoulli, ‘probs’ does not correspond to a probability and ‘logits’ does not correspond to log-odds, but the same names are used due to the similarity with the Bernoulli. See [1] for more details. Example: >>> m = ContinuousBernoulli(torch.tensor([0.3])) >>> m.sample() tensor([ 0.2538]) Parameters probs (Number, Tensor) – (0,1) valued parameters logits (Number, Tensor) – real valued parameters whose sigmoid matches ‘probs’ [1] The continuous Bernoulli: fixing a pervasive error in variational autoencoders, Loaiza-Ganem G and Cunningham JP, NeurIPS 2019. https://arxiv.org/abs/1907.06845 arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] logits [source] property mean property param_shape probs [source] rsample(sample_shape=torch.Size([])) [source] sample(sample_shape=torch.Size([])) [source] property stddev support = Interval(lower_bound=0.0, upper_bound=1.0) property variance Dirichlet class torch.distributions.dirichlet.Dirichlet(concentration, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Dirichlet distribution parameterized by concentration concentration. Example: >>> m = Dirichlet(torch.tensor([0.5, 0.5])) >>> m.sample() # Dirichlet distributed with concentrarion concentration tensor([ 0.1046, 0.8954]) Parameters concentration (Tensor) – concentration parameter of the distribution (often referred to as alpha) arg_constraints = {'concentration': IndependentConstraint(GreaterThan(lower_bound=0.0), 1)} entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=()) [source] support = Simplex() property variance Exponential class torch.distributions.exponential.Exponential(rate, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Exponential distribution parameterized by rate. Example: >>> m = Exponential(torch.tensor([1.0])) >>> m.sample() # Exponential distributed with rate=1 tensor([ 0.1046]) Parameters rate (float or Tensor) – rate = 1 / scale of the distribution arg_constraints = {'rate': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] property stddev support = GreaterThan(lower_bound=0.0) property variance FisherSnedecor class torch.distributions.fishersnedecor.FisherSnedecor(df1, df2, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Fisher-Snedecor distribution parameterized by df1 and df2. Example: >>> m = FisherSnedecor(torch.tensor([1.0]), torch.tensor([2.0])) >>> m.sample() # Fisher-Snedecor-distributed with df1=1 and df2=2 tensor([ 0.2453]) Parameters df1 (float or Tensor) – degrees of freedom parameter 1 df2 (float or Tensor) – degrees of freedom parameter 2 arg_constraints = {'df1': GreaterThan(lower_bound=0.0), 'df2': GreaterThan(lower_bound=0.0)} expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] support = GreaterThan(lower_bound=0.0) property variance Gamma class torch.distributions.gamma.Gamma(concentration, rate, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Gamma distribution parameterized by shape concentration and rate. Example: >>> m = Gamma(torch.tensor([1.0]), torch.tensor([1.0])) >>> m.sample() # Gamma distributed with concentration=1 and rate=1 tensor([ 0.1046]) Parameters concentration (float or Tensor) – shape parameter of the distribution (often referred to as alpha) rate (float or Tensor) – rate = 1 / scale of the distribution (often referred to as beta) arg_constraints = {'concentration': GreaterThan(lower_bound=0.0), 'rate': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] support = GreaterThan(lower_bound=0.0) property variance Geometric class torch.distributions.geometric.Geometric(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Geometric distribution parameterized by probs, where probs is the probability of success of Bernoulli trials. It represents the probability that in k+1k + 1 Bernoulli trials, the first kk trials failed, before seeing a success. Samples are non-negative integers [0, inf⁡\inf ). Example: >>> m = Geometric(torch.tensor([0.3])) >>> m.sample() # underlying Bernoulli has 30% chance 1; 70% chance 0 tensor([ 2.]) Parameters probs (Number, Tensor) – the probability of sampling 1. Must be in range (0, 1] logits (Number, Tensor) – the log-odds of sampling 1. arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} entropy() [source] expand(batch_shape, _instance=None) [source] log_prob(value) [source] logits [source] property mean probs [source] sample(sample_shape=torch.Size([])) [source] support = IntegerGreaterThan(lower_bound=0) property variance Gumbel class torch.distributions.gumbel.Gumbel(loc, scale, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Samples from a Gumbel Distribution. Examples: >>> m = Gumbel(torch.tensor([1.0]), torch.tensor([2.0])) >>> m.sample() # sample from Gumbel distribution with loc=1, scale=2 tensor([ 1.0124]) Parameters loc (float or Tensor) – Location parameter of the distribution scale (float or Tensor) – Scale parameter of the distribution arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] log_prob(value) [source] property mean property stddev support = Real() property variance HalfCauchy class torch.distributions.half_cauchy.HalfCauchy(scale, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Creates a half-Cauchy distribution parameterized by scale where: X ~ Cauchy(0, scale) Y = |X| ~ HalfCauchy(scale) Example: >>> m = HalfCauchy(torch.tensor([1.0])) >>> m.sample() # half-cauchy distributed with scale=1 tensor([ 2.3214]) Parameters scale (float or Tensor) – scale of the full Cauchy distribution arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(prob) [source] log_prob(value) [source] property mean property scale support = GreaterThan(lower_bound=0.0) property variance HalfNormal class torch.distributions.half_normal.HalfNormal(scale, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Creates a half-normal distribution parameterized by scale where: X ~ Normal(0, scale) Y = |X| ~ HalfNormal(scale) Example: >>> m = HalfNormal(torch.tensor([1.0])) >>> m.sample() # half-normal distributed with scale=1 tensor([ 0.1046]) Parameters scale (float or Tensor) – scale of the full Normal distribution arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(prob) [source] log_prob(value) [source] property mean property scale support = GreaterThan(lower_bound=0.0) property variance Independent class torch.distributions.independent.Independent(base_distribution, reinterpreted_batch_ndims, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Reinterprets some of the batch dims of a distribution as event dims. This is mainly useful for changing the shape of the result of log_prob(). For example to create a diagonal Normal distribution with the same shape as a Multivariate Normal distribution (so they are interchangeable), you can: >>> loc = torch.zeros(3) >>> scale = torch.ones(3) >>> mvn = MultivariateNormal(loc, scale_tril=torch.diag(scale)) >>> [mvn.batch_shape, mvn.event_shape] [torch.Size(()), torch.Size((3,))] >>> normal = Normal(loc, scale) >>> [normal.batch_shape, normal.event_shape] [torch.Size((3,)), torch.Size(())] >>> diagn = Independent(normal, 1) >>> [diagn.batch_shape, diagn.event_shape] [torch.Size(()), torch.Size((3,))] Parameters base_distribution (torch.distributions.distribution.Distribution) – a base distribution reinterpreted_batch_ndims (int) – the number of batch dims to reinterpret as event dims arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] property has_enumerate_support property has_rsample log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] sample(sample_shape=torch.Size([])) [source] property support property variance Kumaraswamy class torch.distributions.kumaraswamy.Kumaraswamy(concentration1, concentration0, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Samples from a Kumaraswamy distribution. Example: >>> m = Kumaraswamy(torch.Tensor([1.0]), torch.Tensor([1.0])) >>> m.sample() # sample from a Kumaraswamy distribution with concentration alpha=1 and beta=1 tensor([ 0.1729]) Parameters concentration1 (float or Tensor) – 1st concentration parameter of the distribution (often referred to as alpha) concentration0 (float or Tensor) – 2nd concentration parameter of the distribution (often referred to as beta) arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'concentration0': GreaterThan(lower_bound=0.0), 'concentration1': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True property mean support = Interval(lower_bound=0.0, upper_bound=1.0) property variance LKJCholesky class torch.distributions.lkj_cholesky.LKJCholesky(dim, concentration=1.0, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution LKJ distribution for lower Cholesky factor of correlation matrices. The distribution is controlled by concentration parameter η\eta to make the probability of the correlation matrix MM generated from a Cholesky factor propotional to det⁡(M)η−1\det(M)^{\eta - 1} . Because of that, when concentration == 1, we have a uniform distribution over Cholesky factors of correlation matrices. Note that this distribution samples the Cholesky factor of correlation matrices and not the correlation matrices themselves and thereby differs slightly from the derivations in [1] for the LKJCorr distribution. For sampling, this uses the Onion method from [1] Section 3. L ~ LKJCholesky(dim, concentration) X = L @ L’ ~ LKJCorr(dim, concentration) Example: >>> l = LKJCholesky(3, 0.5) >>> l.sample() # l @ l.T is a sample of a correlation 3x3 matrix tensor([[ 1.0000, 0.0000, 0.0000], [ 0.3516, 0.9361, 0.0000], [-0.1899, 0.4748, 0.8593]]) Parameters dimension (dim) – dimension of the matrices concentration (float or Tensor) – concentration/shape parameter of the distribution (often referred to as eta) References [1] Generating random correlation matrices based on vines and extended onion method, Daniel Lewandowski, Dorota Kurowicka, Harry Joe. arg_constraints = {'concentration': GreaterThan(lower_bound=0.0)} expand(batch_shape, _instance=None) [source] log_prob(value) [source] sample(sample_shape=torch.Size([])) [source] support = CorrCholesky() Laplace class torch.distributions.laplace.Laplace(loc, scale, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Laplace distribution parameterized by loc and scale. Example: >>> m = Laplace(torch.tensor([0.0]), torch.tensor([1.0])) >>> m.sample() # Laplace distributed with loc=0, scale=1 tensor([ 0.1046]) Parameters loc (float or Tensor) – mean of the distribution scale (float or Tensor) – scale of the distribution arg_constraints = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] property stddev support = Real() property variance LogNormal class torch.distributions.log_normal.LogNormal(loc, scale, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Creates a log-normal distribution parameterized by loc and scale where: X ~ Normal(loc, scale) Y = exp(X) ~ LogNormal(loc, scale) Example: >>> m = LogNormal(torch.tensor([0.0]), torch.tensor([1.0])) >>> m.sample() # log-normal distributed with mean=0 and stddev=1 tensor([ 0.1046]) Parameters loc (float or Tensor) – mean of log of distribution scale (float or Tensor) – standard deviation of log of the distribution arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True property loc property mean property scale support = GreaterThan(lower_bound=0.0) property variance LowRankMultivariateNormal class torch.distributions.lowrank_multivariate_normal.LowRankMultivariateNormal(loc, cov_factor, cov_diag, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a multivariate normal distribution with covariance matrix having a low-rank form parameterized by cov_factor and cov_diag: covariance_matrix = cov_factor @ cov_factor.T + cov_diag Example >>> m = LowRankMultivariateNormal(torch.zeros(2), torch.tensor([[1.], [0.]]), torch.ones(2)) >>> m.sample() # normally distributed with mean=`[0,0]`, cov_factor=`[[1],[0]]`, cov_diag=`[1,1]` tensor([-0.2102, -0.5429]) Parameters loc (Tensor) – mean of the distribution with shape batch_shape + event_shape cov_factor (Tensor) – factor part of low-rank form of covariance matrix with shape batch_shape + event_shape + (rank,) cov_diag (Tensor) – diagonal part of low-rank form of covariance matrix with shape batch_shape + event_shape Note The computation for determinant and inverse of covariance matrix is avoided when cov_factor.shape[1] << cov_factor.shape[0] thanks to Woodbury matrix identity and matrix determinant lemma. Thanks to these formulas, we just need to compute the determinant and inverse of the small size “capacitance” matrix: capacitance = I + cov_factor.T @ inv(cov_diag) @ cov_factor arg_constraints = {'cov_diag': IndependentConstraint(GreaterThan(lower_bound=0.0), 1), 'cov_factor': IndependentConstraint(Real(), 2), 'loc': IndependentConstraint(Real(), 1)} covariance_matrix [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean precision_matrix [source] rsample(sample_shape=torch.Size([])) [source] scale_tril [source] support = IndependentConstraint(Real(), 1) variance [source] MixtureSameFamily class torch.distributions.mixture_same_family.MixtureSameFamily(mixture_distribution, component_distribution, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution The MixtureSameFamily distribution implements a (batch of) mixture distribution where all component are from different parameterizations of the same distribution type. It is parameterized by a Categorical “selecting distribution” (over k component) and a component distribution, i.e., a Distribution with a rightmost batch shape (equal to [k]) which indexes each (batch of) component. Examples: # Construct Gaussian Mixture Model in 1D consisting of 5 equally # weighted normal distributions >>> mix = D.Categorical(torch.ones(5,)) >>> comp = D.Normal(torch.randn(5,), torch.rand(5,)) >>> gmm = MixtureSameFamily(mix, comp) # Construct Gaussian Mixture Modle in 2D consisting of 5 equally # weighted bivariate normal distributions >>> mix = D.Categorical(torch.ones(5,)) >>> comp = D.Independent(D.Normal( torch.randn(5,2), torch.rand(5,2)), 1) >>> gmm = MixtureSameFamily(mix, comp) # Construct a batch of 3 Gaussian Mixture Models in 2D each # consisting of 5 random weighted bivariate normal distributions >>> mix = D.Categorical(torch.rand(3,5)) >>> comp = D.Independent(D.Normal( torch.randn(3,5,2), torch.rand(3,5,2)), 1) >>> gmm = MixtureSameFamily(mix, comp) Parameters mixture_distribution – torch.distributions.Categorical-like instance. Manages the probability of selecting component. The number of categories must match the rightmost batch dimension of the component_distribution. Must have either scalar batch_shape or batch_shape matching component_distribution.batch_shape[:-1] component_distribution – torch.distributions.Distribution-like instance. Right-most batch dimension indexes component. arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {} cdf(x) [source] property component_distribution expand(batch_shape, _instance=None) [source] has_rsample = False log_prob(x) [source] property mean property mixture_distribution sample(sample_shape=torch.Size([])) [source] property support property variance Multinomial class torch.distributions.multinomial.Multinomial(total_count=1, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Multinomial distribution parameterized by total_count and either probs or logits (but not both). The innermost dimension of probs indexes over categories. All other dimensions index over batches. Note that total_count need not be specified if only log_prob() is called (see example below) Note The probs argument must be non-negative, finite and have a non-zero sum, and it will be normalized to sum to 1 along the last dimension. attr:probs will return this normalized value. The logits argument will be interpreted as unnormalized log probabilities and can therefore be any real number. It will likewise be normalized so that the resulting probabilities sum to 1 along the last dimension. attr:logits will return this normalized value. sample() requires a single shared total_count for all parameters and samples. log_prob() allows different total_count for each parameter and sample. Example: >>> m = Multinomial(100, torch.tensor([ 1., 1., 1., 1.])) >>> x = m.sample() # equal probability of 0, 1, 2, 3 tensor([ 21., 24., 30., 25.]) >>> Multinomial(probs=torch.tensor([1., 1., 1., 1.])).log_prob(x) tensor([-4.1338]) Parameters total_count (int) – number of trials probs (Tensor) – event probabilities logits (Tensor) – event log probabilities (unnormalized) arg_constraints = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()} expand(batch_shape, _instance=None) [source] log_prob(value) [source] property logits property mean property param_shape property probs sample(sample_shape=torch.Size([])) [source] property support total_count: int = None property variance MultivariateNormal class torch.distributions.multivariate_normal.MultivariateNormal(loc, covariance_matrix=None, precision_matrix=None, scale_tril=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a multivariate normal (also called Gaussian) distribution parameterized by a mean vector and a covariance matrix. The multivariate normal distribution can be parameterized either in terms of a positive definite covariance matrix Σ\mathbf{\Sigma} or a positive definite precision matrix Σ−1\mathbf{\Sigma}^{-1} or a lower-triangular matrix L\mathbf{L} with positive-valued diagonal entries, such that Σ=LL⊤\mathbf{\Sigma} = \mathbf{L}\mathbf{L}^\top . This triangular matrix can be obtained via e.g. Cholesky decomposition of the covariance. Example >>> m = MultivariateNormal(torch.zeros(2), torch.eye(2)) >>> m.sample() # normally distributed with mean=`[0,0]` and covariance_matrix=`I` tensor([-0.2102, -0.5429]) Parameters loc (Tensor) – mean of the distribution covariance_matrix (Tensor) – positive-definite covariance matrix precision_matrix (Tensor) – positive-definite precision matrix scale_tril (Tensor) – lower-triangular factor of covariance, with positive-valued diagonal Note Only one of covariance_matrix or precision_matrix or scale_tril can be specified. Using scale_tril will be more efficient: all computations internally are based on scale_tril. If covariance_matrix or precision_matrix is passed instead, it is only used to compute the corresponding lower triangular matrices using a Cholesky decomposition. arg_constraints = {'covariance_matrix': PositiveDefinite(), 'loc': IndependentConstraint(Real(), 1), 'precision_matrix': PositiveDefinite(), 'scale_tril': LowerCholesky()} covariance_matrix [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean precision_matrix [source] rsample(sample_shape=torch.Size([])) [source] scale_tril [source] support = IndependentConstraint(Real(), 1) property variance NegativeBinomial class torch.distributions.negative_binomial.NegativeBinomial(total_count, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Negative Binomial distribution, i.e. distribution of the number of successful independent and identical Bernoulli trials before total_count failures are achieved. The probability of failure of each Bernoulli trial is probs. Parameters total_count (float or Tensor) – non-negative number of negative Bernoulli trials to stop, although the distribution is still valid for real valued count probs (Tensor) – Event probabilities of failure in the half open interval [0, 1) logits (Tensor) – Event log-odds for probabilities of failure arg_constraints = {'logits': Real(), 'probs': HalfOpenInterval(lower_bound=0.0, upper_bound=1.0), 'total_count': GreaterThanEq(lower_bound=0)} expand(batch_shape, _instance=None) [source] log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] support = IntegerGreaterThan(lower_bound=0) property variance Normal class torch.distributions.normal.Normal(loc, scale, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a normal (also called Gaussian) distribution parameterized by loc and scale. Example: >>> m = Normal(torch.tensor([0.0]), torch.tensor([1.0])) >>> m.sample() # normally distributed with loc=0 and scale=1 tensor([ 0.1046]) Parameters loc (float or Tensor) – mean of the distribution (often referred to as mu) scale (float or Tensor) – standard deviation of the distribution (often referred to as sigma) arg_constraints = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] sample(sample_shape=torch.Size([])) [source] property stddev support = Real() property variance OneHotCategorical class torch.distributions.one_hot_categorical.OneHotCategorical(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a one-hot categorical distribution parameterized by probs or logits. Samples are one-hot coded vectors of size probs.size(-1). Note The probs argument must be non-negative, finite and have a non-zero sum, and it will be normalized to sum to 1 along the last dimension. attr:probs will return this normalized value. The logits argument will be interpreted as unnormalized log probabilities and can therefore be any real number. It will likewise be normalized so that the resulting probabilities sum to 1 along the last dimension. attr:logits will return this normalized value. See also: torch.distributions.Categorical() for specifications of probs and logits. Example: >>> m = OneHotCategorical(torch.tensor([ 0.25, 0.25, 0.25, 0.25 ])) >>> m.sample() # equal probability of 0, 1, 2, 3 tensor([ 0., 0., 0., 1.]) Parameters probs (Tensor) – event probabilities logits (Tensor) – event log probabilities (unnormalized) arg_constraints = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] property logits property mean property param_shape property probs sample(sample_shape=torch.Size([])) [source] support = OneHot() property variance Pareto class torch.distributions.pareto.Pareto(scale, alpha, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Samples from a Pareto Type 1 distribution. Example: >>> m = Pareto(torch.tensor([1.0]), torch.tensor([1.0])) >>> m.sample() # sample from a Pareto distribution with scale=1 and alpha=1 tensor([ 1.5623]) Parameters scale (float or Tensor) – Scale parameter of the distribution alpha (float or Tensor) – Shape parameter of the distribution arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'alpha': GreaterThan(lower_bound=0.0), 'scale': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] property mean property support property variance Poisson class torch.distributions.poisson.Poisson(rate, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Poisson distribution parameterized by rate, the rate parameter. Samples are nonnegative integers, with a pmf given by rateke−ratek!\mathrm{rate}^k \frac{e^{-\mathrm{rate}}}{k!} Example: >>> m = Poisson(torch.tensor([4])) >>> m.sample() tensor([ 3.]) Parameters rate (Number, Tensor) – the rate parameter arg_constraints = {'rate': GreaterThan(lower_bound=0.0)} expand(batch_shape, _instance=None) [source] log_prob(value) [source] property mean sample(sample_shape=torch.Size([])) [source] support = IntegerGreaterThan(lower_bound=0) property variance RelaxedBernoulli class torch.distributions.relaxed_bernoulli.RelaxedBernoulli(temperature, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Creates a RelaxedBernoulli distribution, parametrized by temperature, and either probs or logits (but not both). This is a relaxed version of the Bernoulli distribution, so the values are in (0, 1), and has reparametrizable samples. Example: >>> m = RelaxedBernoulli(torch.tensor([2.2]), torch.tensor([0.1, 0.2, 0.3, 0.99])) >>> m.sample() tensor([ 0.2951, 0.3442, 0.8918, 0.9021]) Parameters temperature (Tensor) – relaxation temperature probs (Number, Tensor) – the probability of sampling 1 logits (Number, Tensor) – the log-odds of sampling 1 arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} expand(batch_shape, _instance=None) [source] has_rsample = True property logits property probs support = Interval(lower_bound=0.0, upper_bound=1.0) property temperature LogitRelaxedBernoulli class torch.distributions.relaxed_bernoulli.LogitRelaxedBernoulli(temperature, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a LogitRelaxedBernoulli distribution parameterized by probs or logits (but not both), which is the logit of a RelaxedBernoulli distribution. Samples are logits of values in (0, 1). See [1] for more details. Parameters temperature (Tensor) – relaxation temperature probs (Number, Tensor) – the probability of sampling 1 logits (Number, Tensor) – the log-odds of sampling 1 [1] The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables (Maddison et al, 2017) [2] Categorical Reparametrization with Gumbel-Softmax (Jang et al, 2017) arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} expand(batch_shape, _instance=None) [source] log_prob(value) [source] logits [source] property param_shape probs [source] rsample(sample_shape=torch.Size([])) [source] support = Real() RelaxedOneHotCategorical class torch.distributions.relaxed_categorical.RelaxedOneHotCategorical(temperature, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Creates a RelaxedOneHotCategorical distribution parametrized by temperature, and either probs or logits. This is a relaxed version of the OneHotCategorical distribution, so its samples are on simplex, and are reparametrizable. Example: >>> m = RelaxedOneHotCategorical(torch.tensor([2.2]), torch.tensor([0.1, 0.2, 0.3, 0.4])) >>> m.sample() tensor([ 0.1294, 0.2324, 0.3859, 0.2523]) Parameters temperature (Tensor) – relaxation temperature probs (Tensor) – event probabilities logits (Tensor) – unnormalized log probability for each event arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()} expand(batch_shape, _instance=None) [source] has_rsample = True property logits property probs support = Simplex() property temperature StudentT class torch.distributions.studentT.StudentT(df, loc=0.0, scale=1.0, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Student’s t-distribution parameterized by degree of freedom df, mean loc and scale scale. Example: >>> m = StudentT(torch.tensor([2.0])) >>> m.sample() # Student's t-distributed with degrees of freedom=2 tensor([ 0.1046]) Parameters df (float or Tensor) – degrees of freedom loc (float or Tensor) – mean of the distribution scale (float or Tensor) – scale of the distribution arg_constraints = {'df': GreaterThan(lower_bound=0.0), 'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] support = Real() property variance TransformedDistribution class torch.distributions.transformed_distribution.TransformedDistribution(base_distribution, transforms, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Extension of the Distribution class, which applies a sequence of Transforms to a base distribution. Let f be the composition of transforms applied: X ~ BaseDistribution Y = f(X) ~ TransformedDistribution(BaseDistribution, f) log p(Y) = log p(X) + log |det (dX/dY)| Note that the .event_shape of a TransformedDistribution is the maximum shape of its base distribution and its transforms, since transforms can introduce correlations among events. An example for the usage of TransformedDistribution would be: # Building a Logistic Distribution # X ~ Uniform(0, 1) # f = a + b * logit(X) # Y ~ f(X) ~ Logistic(a, b) base_distribution = Uniform(0, 1) transforms = [SigmoidTransform().inv, AffineTransform(loc=a, scale=b)] logistic = TransformedDistribution(base_distribution, transforms) For more examples, please look at the implementations of Gumbel, HalfCauchy, HalfNormal, LogNormal, Pareto, Weibull, RelaxedBernoulli and RelaxedOneHotCategorical arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {} cdf(value) [source] Computes the cumulative distribution function by inverting the transform(s) and computing the score of the base distribution. expand(batch_shape, _instance=None) [source] property has_rsample icdf(value) [source] Computes the inverse cumulative distribution function using transform(s) and computing the score of the base distribution. log_prob(value) [source] Scores the sample by inverting the transform(s) and computing the score using the score of the base distribution and the log abs det jacobian. rsample(sample_shape=torch.Size([])) [source] Generates a sample_shape shaped reparameterized sample or sample_shape shaped batch of reparameterized samples if the distribution parameters are batched. Samples first from base distribution and applies transform() for every transform in the list. sample(sample_shape=torch.Size([])) [source] Generates a sample_shape shaped sample or sample_shape shaped batch of samples if the distribution parameters are batched. Samples first from base distribution and applies transform() for every transform in the list. property support Uniform class torch.distributions.uniform.Uniform(low, high, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Generates uniformly distributed random samples from the half-open interval [low, high). Example: >>> m = Uniform(torch.tensor([0.0]), torch.tensor([5.0])) >>> m.sample() # uniformly distributed in the range [0.0, 5.0) tensor([ 2.3418]) Parameters low (float or Tensor) – lower range (inclusive). high (float or Tensor) – upper range (exclusive). arg_constraints = {'high': Dependent(), 'low': Dependent()} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] property stddev property support property variance VonMises class torch.distributions.von_mises.VonMises(loc, concentration, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution A circular von Mises distribution. This implementation uses polar coordinates. The loc and value args can be any real number (to facilitate unconstrained optimization), but are interpreted as angles modulo 2 pi. Example:: >>> m = dist.VonMises(torch.tensor([1.0]), torch.tensor([1.0])) >>> m.sample() # von Mises distributed with loc=1 and concentration=1 tensor([1.9777]) Parameters loc (torch.Tensor) – an angle in radians. concentration (torch.Tensor) – concentration parameter arg_constraints = {'concentration': GreaterThan(lower_bound=0.0), 'loc': Real()} expand(batch_shape) [source] has_rsample = False log_prob(value) [source] property mean The provided mean is the circular one. sample(sample_shape=torch.Size([])) [source] The sampling algorithm for the von Mises distribution is based on the following paper: Best, D. J., and Nicholas I. Fisher. “Efficient simulation of the von Mises distribution.” Applied Statistics (1979): 152-157. support = Real() variance [source] The provided variance is the circular one. Weibull class torch.distributions.weibull.Weibull(scale, concentration, validate_args=None) [source] Bases: torch.distributions.transformed_distribution.TransformedDistribution Samples from a two-parameter Weibull distribution. Example >>> m = Weibull(torch.tensor([1.0]), torch.tensor([1.0])) >>> m.sample() # sample from a Weibull distribution with scale=1, concentration=1 tensor([ 0.4784]) Parameters scale (float or Tensor) – Scale parameter of distribution (lambda). concentration (float or Tensor) – Concentration parameter of distribution (k/shape). arg_constraints: Dict[str, torch.distributions.constraints.Constraint] = {'concentration': GreaterThan(lower_bound=0.0), 'scale': GreaterThan(lower_bound=0.0)} entropy() [source] expand(batch_shape, _instance=None) [source] property mean support = GreaterThan(lower_bound=0.0) property variance KL Divergence torch.distributions.kl.kl_divergence(p, q) [source] Compute Kullback-Leibler divergence KL(p∥q)KL(p \| q) between two distributions. KL(p∥q)=∫p(x)log⁡p(x)q(x)dxKL(p \| q) = \int p(x) \log\frac {p(x)} {q(x)} \,dx Parameters p (Distribution) – A Distribution object. q (Distribution) – A Distribution object. Returns A batch of KL divergences of shape batch_shape. Return type Tensor Raises NotImplementedError – If the distribution types have not been registered via register_kl(). torch.distributions.kl.register_kl(type_p, type_q) [source] Decorator to register a pairwise function with kl_divergence(). Usage: @register_kl(Normal, Normal) def kl_normal_normal(p, q): # insert implementation here Lookup returns the most specific (type,type) match ordered by subclass. If the match is ambiguous, a RuntimeWarning is raised. For example to resolve the ambiguous situation: @register_kl(BaseP, DerivedQ) def kl_version1(p, q): ... @register_kl(DerivedP, BaseQ) def kl_version2(p, q): ... you should register a third most-specific implementation, e.g.: register_kl(DerivedP, DerivedQ)(kl_version1) # Break the tie. Parameters type_p (type) – A subclass of Distribution. type_q (type) – A subclass of Distribution. Transforms class torch.distributions.transforms.Transform(cache_size=0) [source] Abstract class for invertable transformations with computable log det jacobians. They are primarily used in torch.distributions.TransformedDistribution. Caching is useful for transforms whose inverses are either expensive or numerically unstable. Note that care must be taken with memoized values since the autograd graph may be reversed. For example while the following works with or without caching: y = t(x) t.log_abs_det_jacobian(x, y).backward() # x will receive gradients. However the following will error when caching due to dependency reversal: y = t(x) z = t.inv(y) grad(z.sum(), [y]) # error because z is x Derived classes should implement one or both of _call() or _inverse(). Derived classes that set bijective=True should also implement log_abs_det_jacobian(). Parameters cache_size (int) – Size of cache. If zero, no caching is done. If one, the latest single value is cached. Only 0 and 1 are supported. Variables ~Transform.domain (Constraint) – The constraint representing valid inputs to this transform. ~Transform.codomain (Constraint) – The constraint representing valid outputs to this transform which are inputs to the inverse transform. ~Transform.bijective (bool) – Whether this transform is bijective. A transform t is bijective iff t.inv(t(x)) == x and t(t.inv(y)) == y for every x in the domain and y in the codomain. Transforms that are not bijective should at least maintain the weaker pseudoinverse properties t(t.inv(t(x)) == t(x) and t.inv(t(t.inv(y))) == t.inv(y). ~Transform.sign (int or Tensor) – For bijective univariate transforms, this should be +1 or -1 depending on whether transform is monotone increasing or decreasing. property inv Returns the inverse Transform of this transform. This should satisfy t.inv.inv is t. property sign Returns the sign of the determinant of the Jacobian, if applicable. In general this only makes sense for bijective transforms. log_abs_det_jacobian(x, y) [source] Computes the log det jacobian log |dy/dx| given input and output. forward_shape(shape) [source] Infers the shape of the forward computation, given the input shape. Defaults to preserving shape. inverse_shape(shape) [source] Infers the shapes of the inverse computation, given the output shape. Defaults to preserving shape. class torch.distributions.transforms.ComposeTransform(parts, cache_size=0) [source] Composes multiple transforms in a chain. The transforms being composed are responsible for caching. Parameters parts (list of Transform) – A list of transforms to compose. cache_size (int) – Size of cache. If zero, no caching is done. If one, the latest single value is cached. Only 0 and 1 are supported. class torch.distributions.transforms.IndependentTransform(base_transform, reinterpreted_batch_ndims, cache_size=0) [source] Wrapper around another transform to treat reinterpreted_batch_ndims-many extra of the right most dimensions as dependent. This has no effect on the forward or backward transforms, but does sum out reinterpreted_batch_ndims-many of the rightmost dimensions in log_abs_det_jacobian(). Parameters base_transform (Transform) – A base transform. reinterpreted_batch_ndims (int) – The number of extra rightmost dimensions to treat as dependent. class torch.distributions.transforms.ReshapeTransform(in_shape, out_shape, cache_size=0) [source] Unit Jacobian transform to reshape the rightmost part of a tensor. Note that in_shape and out_shape must have the same number of elements, just as for torch.Tensor.reshape(). Parameters in_shape (torch.Size) – The input event shape. out_shape (torch.Size) – The output event shape. class torch.distributions.transforms.ExpTransform(cache_size=0) [source] Transform via the mapping y=exp⁡(x)y = \exp(x) . class torch.distributions.transforms.PowerTransform(exponent, cache_size=0) [source] Transform via the mapping y=xexponenty = x^{\text{exponent}} . class torch.distributions.transforms.SigmoidTransform(cache_size=0) [source] Transform via the mapping y=11+exp⁡(−x)y = \frac{1}{1 + \exp(-x)} and x=logit(y)x = \text{logit}(y) . class torch.distributions.transforms.TanhTransform(cache_size=0) [source] Transform via the mapping y=tanh⁡(x)y = \tanh(x) . It is equivalent to ` ComposeTransform([AffineTransform(0., 2.), SigmoidTransform(), AffineTransform(-1., 2.)]) ` However this might not be numerically stable, thus it is recommended to use TanhTransform instead. Note that one should use cache_size=1 when it comes to NaN/Inf values. class torch.distributions.transforms.AbsTransform(cache_size=0) [source] Transform via the mapping y=∣x∣y = |x| . class torch.distributions.transforms.AffineTransform(loc, scale, event_dim=0, cache_size=0) [source] Transform via the pointwise affine mapping y=loc+scale×xy = \text{loc} + \text{scale} \times x . Parameters loc (Tensor or float) – Location parameter. scale (Tensor or float) – Scale parameter. event_dim (int) – Optional size of event_shape. This should be zero for univariate random variables, 1 for distributions over vectors, 2 for distributions over matrices, etc. class torch.distributions.transforms.CorrCholeskyTransform(cache_size=0) [source] Transforms an uncontrained real vector xx with length D∗(D−1)/2D*(D-1)/2 into the Cholesky factor of a D-dimension correlation matrix. This Cholesky factor is a lower triangular matrix with positive diagonals and unit Euclidean norm for each row. The transform is processed as follows: First we convert x into a lower triangular matrix in row order. For each row XiX_i of the lower triangular part, we apply a signed version of class StickBreakingTransform to transform XiX_i into a unit Euclidean length vector using the following steps: - Scales into the interval (−1,1)(-1, 1) domain: ri=tanh⁡(Xi)r_i = \tanh(X_i) . - Transforms into an unsigned domain: zi=ri2z_i = r_i^2 . - Applies si=StickBreakingTransform(zi)s_i = StickBreakingTransform(z_i) . - Transforms back into signed domain: yi=sign(ri)∗siy_i = sign(r_i) * \sqrt{s_i} . class torch.distributions.transforms.SoftmaxTransform(cache_size=0) [source] Transform from unconstrained space to the simplex via y=exp⁡(x)y = \exp(x) then normalizing. This is not bijective and cannot be used for HMC. However this acts mostly coordinate-wise (except for the final normalization), and thus is appropriate for coordinate-wise optimization algorithms. class torch.distributions.transforms.StickBreakingTransform(cache_size=0) [source] Transform from unconstrained space to the simplex of one additional dimension via a stick-breaking process. This transform arises as an iterated sigmoid transform in a stick-breaking construction of the Dirichlet distribution: the first logit is transformed via sigmoid to the first probability and the probability of everything else, and then the process recurses. This is bijective and appropriate for use in HMC; however it mixes coordinates together and is less appropriate for optimization. class torch.distributions.transforms.LowerCholeskyTransform(cache_size=0) [source] Transform from unconstrained matrices to lower-triangular matrices with nonnegative diagonal entries. This is useful for parameterizing positive definite matrices in terms of their Cholesky factorization. class torch.distributions.transforms.StackTransform(tseq, dim=0, cache_size=0) [source] Transform functor that applies a sequence of transforms tseq component-wise to each submatrix at dim in a way compatible with torch.stack(). Example:: x = torch.stack([torch.range(1, 10), torch.range(1, 10)], dim=1) t = StackTransform([ExpTransform(), identity_transform], dim=1) y = t(x) Constraints The following constraints are implemented: constraints.boolean constraints.cat constraints.corr_cholesky constraints.dependent constraints.greater_than(lower_bound) constraints.greater_than_eq(lower_bound) constraints.independent(constraint, reinterpreted_batch_ndims) constraints.integer_interval(lower_bound, upper_bound) constraints.interval(lower_bound, upper_bound) constraints.less_than(upper_bound) constraints.lower_cholesky constraints.lower_triangular constraints.multinomial constraints.nonnegative_integer constraints.one_hot constraints.positive_definite constraints.positive_integer constraints.positive constraints.real_vector constraints.real constraints.simplex constraints.stack constraints.unit_interval class torch.distributions.constraints.Constraint [source] Abstract base class for constraints. A constraint object represents a region over which a variable is valid, e.g. within which a variable can be optimized. Variables ~Constraint.is_discrete (bool) – Whether constrained space is discrete. Defaults to False. ~Constraint.event_dim (int) – Number of rightmost dimensions that together define an event. The check() method will remove this many dimensions when computing validity. check(value) [source] Returns a byte tensor of sample_shape + batch_shape indicating whether each event in value satisfies this constraint. torch.distributions.constraints.dependent_property alias of torch.distributions.constraints._DependentProperty torch.distributions.constraints.independent alias of torch.distributions.constraints._IndependentConstraint torch.distributions.constraints.integer_interval alias of torch.distributions.constraints._IntegerInterval torch.distributions.constraints.greater_than alias of torch.distributions.constraints._GreaterThan torch.distributions.constraints.greater_than_eq alias of torch.distributions.constraints._GreaterThanEq torch.distributions.constraints.less_than alias of torch.distributions.constraints._LessThan torch.distributions.constraints.multinomial alias of torch.distributions.constraints._Multinomial torch.distributions.constraints.interval alias of torch.distributions.constraints._Interval torch.distributions.constraints.half_open_interval alias of torch.distributions.constraints._HalfOpenInterval torch.distributions.constraints.cat alias of torch.distributions.constraints._Cat torch.distributions.constraints.stack alias of torch.distributions.constraints._Stack Constraint Registry PyTorch provides two global ConstraintRegistry objects that link Constraint objects to Transform objects. These objects both input constraints and return transforms, but they have different guarantees on bijectivity. biject_to(constraint) looks up a bijective Transform from constraints.real to the given constraint. The returned transform is guaranteed to have .bijective = True and should implement .log_abs_det_jacobian(). transform_to(constraint) looks up a not-necessarily bijective Transform from constraints.real to the given constraint. The returned transform is not guaranteed to implement .log_abs_det_jacobian(). The transform_to() registry is useful for performing unconstrained optimization on constrained parameters of probability distributions, which are indicated by each distribution’s .arg_constraints dict. These transforms often overparameterize a space in order to avoid rotation; they are thus more suitable for coordinate-wise optimization algorithms like Adam: loc = torch.zeros(100, requires_grad=True) unconstrained = torch.zeros(100, requires_grad=True) scale = transform_to(Normal.arg_constraints['scale'])(unconstrained) loss = -Normal(loc, scale).log_prob(data).sum() The biject_to() registry is useful for Hamiltonian Monte Carlo, where samples from a probability distribution with constrained .support are propagated in an unconstrained space, and algorithms are typically rotation invariant.: dist = Exponential(rate) unconstrained = torch.zeros(100, requires_grad=True) sample = biject_to(dist.support)(unconstrained) potential_energy = -dist.log_prob(sample).sum() Note An example where transform_to and biject_to differ is constraints.simplex: transform_to(constraints.simplex) returns a SoftmaxTransform that simply exponentiates and normalizes its inputs; this is a cheap and mostly coordinate-wise operation appropriate for algorithms like SVI. In contrast, biject_to(constraints.simplex) returns a StickBreakingTransform that bijects its input down to a one-fewer-dimensional space; this a more expensive less numerically stable transform but is needed for algorithms like HMC. The biject_to and transform_to objects can be extended by user-defined constraints and transforms using their .register() method either as a function on singleton constraints: transform_to.register(my_constraint, my_transform) or as a decorator on parameterized constraints: @transform_to.register(MyConstraintClass) def my_factory(constraint): assert isinstance(constraint, MyConstraintClass) return MyTransform(constraint.param1, constraint.param2) You can create your own registry by creating a new ConstraintRegistry object. class torch.distributions.constraint_registry.ConstraintRegistry [source] Registry to link constraints to transforms. register(constraint, factory=None) [source] Registers a Constraint subclass in this registry. Usage: @my_registry.register(MyConstraintClass) def construct_transform(constraint): assert isinstance(constraint, MyConstraint) return MyTransform(constraint.arg_constraints) Parameters constraint (subclass of Constraint) – A subclass of Constraint, or a singleton object of the desired class. factory (callable) – A callable that inputs a constraint object and returns a Transform object.
torch.distributions
class torch.distributions.bernoulli.Bernoulli(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Creates a Bernoulli distribution parameterized by probs or logits (but not both). Samples are binary (0 or 1). They take the value 1 with probability p and 0 with probability 1 - p. Example: >>> m = Bernoulli(torch.tensor([0.3])) >>> m.sample() # 30% chance 1; 70% chance 0 tensor([ 0.]) Parameters probs (Number, Tensor) – the probability of sampling 1 logits (Number, Tensor) – the log-odds of sampling 1 arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] support = Boolean() property variance
torch.distributions#torch.distributions.bernoulli.Bernoulli
arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0)}
torch.distributions#torch.distributions.bernoulli.Bernoulli.arg_constraints
entropy() [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.entropy
enumerate_support(expand=True) [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.enumerate_support
expand(batch_shape, _instance=None) [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.expand
has_enumerate_support = True
torch.distributions#torch.distributions.bernoulli.Bernoulli.has_enumerate_support
logits [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.logits
log_prob(value) [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.log_prob
property mean
torch.distributions#torch.distributions.bernoulli.Bernoulli.mean
property param_shape
torch.distributions#torch.distributions.bernoulli.Bernoulli.param_shape
probs [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.probs
sample(sample_shape=torch.Size([])) [source]
torch.distributions#torch.distributions.bernoulli.Bernoulli.sample
support = Boolean()
torch.distributions#torch.distributions.bernoulli.Bernoulli.support
property variance
torch.distributions#torch.distributions.bernoulli.Bernoulli.variance
class torch.distributions.beta.Beta(concentration1, concentration0, validate_args=None) [source] Bases: torch.distributions.exp_family.ExponentialFamily Beta distribution parameterized by concentration1 and concentration0. Example: >>> m = Beta(torch.tensor([0.5]), torch.tensor([0.5])) >>> m.sample() # Beta distributed with concentration concentration1 and concentration0 tensor([ 0.1046]) Parameters concentration1 (float or Tensor) – 1st concentration parameter of the distribution (often referred to as alpha) concentration0 (float or Tensor) – 2nd concentration parameter of the distribution (often referred to as beta) arg_constraints = {'concentration0': GreaterThan(lower_bound=0.0), 'concentration1': GreaterThan(lower_bound=0.0)} property concentration0 property concentration1 entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True log_prob(value) [source] property mean rsample(sample_shape=()) [source] support = Interval(lower_bound=0.0, upper_bound=1.0) property variance
torch.distributions#torch.distributions.beta.Beta
arg_constraints = {'concentration0': GreaterThan(lower_bound=0.0), 'concentration1': GreaterThan(lower_bound=0.0)}
torch.distributions#torch.distributions.beta.Beta.arg_constraints
property concentration0
torch.distributions#torch.distributions.beta.Beta.concentration0
property concentration1
torch.distributions#torch.distributions.beta.Beta.concentration1
entropy() [source]
torch.distributions#torch.distributions.beta.Beta.entropy
expand(batch_shape, _instance=None) [source]
torch.distributions#torch.distributions.beta.Beta.expand
has_rsample = True
torch.distributions#torch.distributions.beta.Beta.has_rsample
log_prob(value) [source]
torch.distributions#torch.distributions.beta.Beta.log_prob
property mean
torch.distributions#torch.distributions.beta.Beta.mean
rsample(sample_shape=()) [source]
torch.distributions#torch.distributions.beta.Beta.rsample
support = Interval(lower_bound=0.0, upper_bound=1.0)
torch.distributions#torch.distributions.beta.Beta.support
property variance
torch.distributions#torch.distributions.beta.Beta.variance
class torch.distributions.binomial.Binomial(total_count=1, probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a Binomial distribution parameterized by total_count and either probs or logits (but not both). total_count must be broadcastable with probs/logits. Example: >>> m = Binomial(100, torch.tensor([0 , .2, .8, 1])) >>> x = m.sample() tensor([ 0., 22., 71., 100.]) >>> m = Binomial(torch.tensor([[5.], [10.]]), torch.tensor([0.5, 0.8])) >>> x = m.sample() tensor([[ 4., 5.], [ 7., 6.]]) Parameters total_count (int or Tensor) – number of Bernoulli trials probs (Tensor) – Event probabilities logits (Tensor) – Event log-odds arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0), 'total_count': IntegerGreaterThan(lower_bound=0)} enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] property support property variance
torch.distributions#torch.distributions.binomial.Binomial
arg_constraints = {'logits': Real(), 'probs': Interval(lower_bound=0.0, upper_bound=1.0), 'total_count': IntegerGreaterThan(lower_bound=0)}
torch.distributions#torch.distributions.binomial.Binomial.arg_constraints
enumerate_support(expand=True) [source]
torch.distributions#torch.distributions.binomial.Binomial.enumerate_support
expand(batch_shape, _instance=None) [source]
torch.distributions#torch.distributions.binomial.Binomial.expand
has_enumerate_support = True
torch.distributions#torch.distributions.binomial.Binomial.has_enumerate_support
logits [source]
torch.distributions#torch.distributions.binomial.Binomial.logits
log_prob(value) [source]
torch.distributions#torch.distributions.binomial.Binomial.log_prob
property mean
torch.distributions#torch.distributions.binomial.Binomial.mean
property param_shape
torch.distributions#torch.distributions.binomial.Binomial.param_shape
probs [source]
torch.distributions#torch.distributions.binomial.Binomial.probs
sample(sample_shape=torch.Size([])) [source]
torch.distributions#torch.distributions.binomial.Binomial.sample
property support
torch.distributions#torch.distributions.binomial.Binomial.support
property variance
torch.distributions#torch.distributions.binomial.Binomial.variance
class torch.distributions.categorical.Categorical(probs=None, logits=None, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Creates a categorical distribution parameterized by either probs or logits (but not both). Note It is equivalent to the distribution that torch.multinomial() samples from. Samples are integers from {0,…,K−1}\{0, \ldots, K-1\} where K is probs.size(-1). If probs is 1-dimensional with length-K, each element is the relative probability of sampling the class at that index. If probs is N-dimensional, the first N-1 dimensions are treated as a batch of relative probability vectors. Note The probs argument must be non-negative, finite and have a non-zero sum, and it will be normalized to sum to 1 along the last dimension. attr:probs will return this normalized value. The logits argument will be interpreted as unnormalized log probabilities and can therefore be any real number. It will likewise be normalized so that the resulting probabilities sum to 1 along the last dimension. attr:logits will return this normalized value. See also: torch.multinomial() Example: >>> m = Categorical(torch.tensor([ 0.25, 0.25, 0.25, 0.25 ])) >>> m.sample() # equal probability of 0, 1, 2, 3 tensor(3) Parameters probs (Tensor) – event probabilities logits (Tensor) – event log probabilities (unnormalized) arg_constraints = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()} entropy() [source] enumerate_support(expand=True) [source] expand(batch_shape, _instance=None) [source] has_enumerate_support = True log_prob(value) [source] logits [source] property mean property param_shape probs [source] sample(sample_shape=torch.Size([])) [source] property support property variance
torch.distributions#torch.distributions.categorical.Categorical
arg_constraints = {'logits': IndependentConstraint(Real(), 1), 'probs': Simplex()}
torch.distributions#torch.distributions.categorical.Categorical.arg_constraints
entropy() [source]
torch.distributions#torch.distributions.categorical.Categorical.entropy
enumerate_support(expand=True) [source]
torch.distributions#torch.distributions.categorical.Categorical.enumerate_support
expand(batch_shape, _instance=None) [source]
torch.distributions#torch.distributions.categorical.Categorical.expand
has_enumerate_support = True
torch.distributions#torch.distributions.categorical.Categorical.has_enumerate_support
logits [source]
torch.distributions#torch.distributions.categorical.Categorical.logits
log_prob(value) [source]
torch.distributions#torch.distributions.categorical.Categorical.log_prob
property mean
torch.distributions#torch.distributions.categorical.Categorical.mean
property param_shape
torch.distributions#torch.distributions.categorical.Categorical.param_shape
probs [source]
torch.distributions#torch.distributions.categorical.Categorical.probs
sample(sample_shape=torch.Size([])) [source]
torch.distributions#torch.distributions.categorical.Categorical.sample
property support
torch.distributions#torch.distributions.categorical.Categorical.support
property variance
torch.distributions#torch.distributions.categorical.Categorical.variance
class torch.distributions.cauchy.Cauchy(loc, scale, validate_args=None) [source] Bases: torch.distributions.distribution.Distribution Samples from a Cauchy (Lorentz) distribution. The distribution of the ratio of independent normally distributed random variables with means 0 follows a Cauchy distribution. Example: >>> m = Cauchy(torch.tensor([0.0]), torch.tensor([1.0])) >>> m.sample() # sample from a Cauchy distribution with loc=0 and scale=1 tensor([ 2.3214]) Parameters loc (float or Tensor) – mode or median of the distribution. scale (float or Tensor) – half width at half maximum. arg_constraints = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)} cdf(value) [source] entropy() [source] expand(batch_shape, _instance=None) [source] has_rsample = True icdf(value) [source] log_prob(value) [source] property mean rsample(sample_shape=torch.Size([])) [source] support = Real() property variance
torch.distributions#torch.distributions.cauchy.Cauchy
arg_constraints = {'loc': Real(), 'scale': GreaterThan(lower_bound=0.0)}
torch.distributions#torch.distributions.cauchy.Cauchy.arg_constraints
cdf(value) [source]
torch.distributions#torch.distributions.cauchy.Cauchy.cdf
entropy() [source]
torch.distributions#torch.distributions.cauchy.Cauchy.entropy
expand(batch_shape, _instance=None) [source]
torch.distributions#torch.distributions.cauchy.Cauchy.expand
has_rsample = True
torch.distributions#torch.distributions.cauchy.Cauchy.has_rsample
icdf(value) [source]
torch.distributions#torch.distributions.cauchy.Cauchy.icdf
log_prob(value) [source]
torch.distributions#torch.distributions.cauchy.Cauchy.log_prob
property mean
torch.distributions#torch.distributions.cauchy.Cauchy.mean
rsample(sample_shape=torch.Size([])) [source]
torch.distributions#torch.distributions.cauchy.Cauchy.rsample