python_code
stringlengths
0
229k
#!/usr/bin/env python3 import copy import random import unittest from typing import Callable import numpy as np import torch from captum.log import patch_methods def deep_copy_args(func: Callable): def copy_args(*args, **kwargs): return func( *(copy.deepcopy(x) for x in args), **{...
#! /usr/bin/env python3 # Configuration file for the Sphinx documentation builder. # # This file does only contain a selection of the most common options. For a # full list see the documentation: # http://www.sphinx-doc.org/en/master/config # -- Path setup ------------------------------------------------------------...
#!/usr/bin/env python3 import argparse import json import os import nbformat from bs4 import BeautifulSoup from nbconvert import HTMLExporter, ScriptExporter TEMPLATE = """const CWD = process.cwd(); const React = require('react'); const Tutorial = require(`${{CWD}}/core/Tutorial.js`); class TutorialPage extends Re...
#!/usr/bin/env python3 import argparse import json from bs4 import BeautifulSoup BASE_URL = "/" def updateVersionHTML(base_path, base_url=BASE_URL): with open(base_path + "/captum-master/website/_versions.json", "rb") as infile: versions = json.loads(infile.read()) with open(base_path + "/new-site...
#!/usr/bin/env python3 import argparse import os from bs4 import BeautifulSoup # no need to import css from built path # coz docusaurus merge all css files within static folder automatically # https://v1.docusaurus.io/docs/en/api-pages#styles base_scripts = """ <script type="text/javascript" id="documentation_option...
#!/usr/bin/env python3 import captum.attr as attr # noqa import captum.concept as concept # noqa import captum.influence as influence # noqa import captum.log as log # noqa import captum.metrics as metrics # noqa import captum.robust as robust # noqa __version__ = "0.6.0"
#!/usr/bin/env python3 from captum.metrics._core.infidelity import ( # noqa infidelity, infidelity_perturb_func_decorator, ) from captum.metrics._core.sensitivity import sensitivity_max # noqa
#!/usr/bin/env python3 import warnings from typing import Callable, Tuple import torch from torch import Tensor def _divide_and_aggregate_metrics( inputs: Tuple[Tensor, ...], n_perturb_samples: int, metric_func: Callable, agg_func: Callable = torch.add, max_examples_per_batch: int = None, ) -> T...
#!/usr/bin/env python3
#!/usr/bin/env python3 from typing import Any, Callable, cast, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_baseline, _format_tensor_into_tuples, _run_forward, ExpansionTypes, safe...
#!/usr/bin/env python3 from copy import deepcopy from inspect import signature from typing import Any, Callable, cast, Tuple, Union import torch from captum._utils.common import ( _expand_and_update_additional_forward_args, _expand_and_update_baselines, _expand_and_update_target, _format_baseline, ...
#!/usr/bin/env python3 import threading import typing import warnings from collections import defaultdict from typing import Any, Callable, cast, Dict, List, Optional, Sequence, Tuple, Union import torch from captum._utils.common import ( _reduce_list, _run_forward, _sort_key_list, _verify_select_neuro...
from collections import defaultdict from enum import Enum from typing import cast, DefaultDict, Iterable, List, Optional, Tuple, Union import torch from captum._utils.common import _format_tensor_into_tuples, _register_backward_hook from torch import Tensor from torch.nn import Module def _reset_sample_grads(module:...
#!/usr/bin/env python3 import typing from enum import Enum from functools import reduce from inspect import signature from typing import Any, Callable, cast, Dict, List, overload, Tuple, Union import numpy as np import torch from captum._utils.typing import ( BaselineType, Literal, TargetType, TensorOr...
#!/usr/bin/env python3 import glob import os import re import warnings from typing import Any, List, Optional, Tuple, Union import captum._utils.common as common import torch from captum.attr import LayerActivation from torch import Tensor from torch.nn import Module from torch.utils.data import DataLoader, Dataset ...
#!/usr/bin/env python3 import sys import warnings from time import time from typing import cast, Iterable, Sized, TextIO from captum._utils.typing import Literal try: from tqdm.auto import tqdm except ImportError: tqdm = None class DisableErrorIOWrapper(object): def __init__(self, wrapped: TextIO) -> N...
#!/usr/bin/env python3 from typing import List, Tuple, TYPE_CHECKING, TypeVar, Union from torch import Tensor from torch.nn import Module if TYPE_CHECKING: import sys if sys.version_info >= (3, 8): from typing import Literal # noqa: F401 else: from typing_extensions import Literal # no...
from captum._utils.models.linear_model import ( LinearModel, SGDLasso, SGDLinearModel, SGDLinearRegression, SGDRidge, SkLearnLasso, SkLearnLinearModel, SkLearnLinearRegression, SkLearnRidge, ) from captum._utils.models.model import Model __all__ = [ "Model", "LinearModel", ...
#!/usr/bin/env python3 from abc import ABC, abstractmethod from typing import Dict, Optional, Union from captum._utils.typing import TensorOrTupleOfTensorsGeneric from torch import Tensor from torch.utils.data import DataLoader class Model(ABC): r""" Abstract Class to describe the interface of a trainable m...
from captum._utils.models.linear_model.model import ( LinearModel, SGDLasso, SGDLinearModel, SGDLinearRegression, SGDRidge, SkLearnLasso, SkLearnLinearModel, SkLearnLinearRegression, SkLearnRidge, ) __all__ = [ "LinearModel", "SGDLinearModel", "SGDLasso", "SGDRidge",...
from typing import Callable, cast, List, Optional import torch.nn as nn from captum._utils.models.model import Model from torch import Tensor from torch.utils.data import DataLoader class LinearModel(nn.Module, Model): SUPPORTED_NORMS: List[Optional[str]] = [None, "batch_norm", "layer_norm"] def __init__(se...
import time import warnings from typing import Any, Callable, Dict, List, Optional import torch import torch.nn as nn from captum._utils.models.linear_model.model import LinearModel from torch.utils.data import DataLoader def l2_loss(x1, x2, weights=None): if weights is None: return torch.mean((x1 - x2) ...
from captum.insights.attr_vis import AttributionVisualizer, Batch, features # noqa
# for legacy purposes import warnings from captum.insights.attr_vis.example import * # noqa warnings.warn( "Deprecated. Please import from captum.insights.attr_vis.example instead." ) main() # noqa
#!/usr/bin/env python3 import logging import os import socket import threading from time import sleep from typing import Optional from captum.log import log_usage from flask import Flask, jsonify, render_template, request from flask_compress import Compress from torch import Tensor app = Flask( __name__, static_f...
#!/usr/bin/env python3 from typing import Any, Callable, Dict, List, NamedTuple, Optional, Tuple, Union from captum.attr import ( Deconvolution, DeepLift, FeatureAblation, GuidedBackprop, InputXGradient, IntegratedGradients, Occlusion, Saliency, ) from captum.attr._utils.approximation_m...
from captum.insights.attr_vis.app import AttributionVisualizer, Batch # noqa
#!/usr/bin/env python3 import base64 import warnings from collections import namedtuple from io import BytesIO from typing import Callable, List, Optional, Union from captum._utils.common import safe_div from captum.attr._utils import visualization as viz from captum.insights.attr_vis._utils.transforms import format_t...
#!/usr/bin/env python3 import inspect from collections import namedtuple from typing import ( Callable, cast, Dict, Iterable, List, Optional, Sequence, Tuple, Union, ) import torch from captum._utils.common import _run_forward, safe_div from captum.insights.attr_vis.config import ( ...
#!/usr/bin/env python3 import os import torch import torch.nn as nn import torchvision import torchvision.transforms as transforms from captum.insights import AttributionVisualizer, Batch from captum.insights.attr_vis.features import ImageFeature def get_classes(): classes = [ "Plane", "Car", ...
#!/usr/bin/env python3 from collections import namedtuple from itertools import cycle from typing import ( Any, Callable, Dict, Iterable, List, NamedTuple, Optional, Tuple, Union, ) import torch from captum.attr import IntegratedGradients from captum.attr._utils.batching import _bat...
#!/usr/bin/env python3 from typing import Callable, List, Optional, Union def format_transforms( transforms: Optional[Union[Callable, List[Callable]]] ) -> List[Callable]: if transforms is None: return [] if callable(transforms): return [transforms] return transforms
#!/usr/bin/env python3 import ipywidgets as widgets from captum.insights import AttributionVisualizer from captum.insights.attr_vis.server import namedtuple_to_dict from traitlets import Dict, Instance, List, observe, Unicode @widgets.register class CaptumInsights(widgets.DOMWidget): """A widget for interacting w...
version_info = (0, 1, 0, "alpha", 0) _specifier_ = {"alpha": "a", "beta": "b", "candidate": "rc", "final": ""} __version__ = "%s.%s.%s%s" % ( version_info[0], version_info[1], version_info[2], "" if version_info[3] == "final" else _specifier_[version_info[3]] + str(version_info[4]), )
from captum.insights.attr_vis.widget._version import __version__, version_info # noqa from captum.insights.attr_vis.widget.widget import * # noqa def _jupyter_nbextension_paths(): return [ { "section": "notebook", "src": "static", "dest": "jupyter-captum-insights", ...
#!/usr/bin/env python3 from captum.robust._core.fgsm import FGSM # noqa from captum.robust._core.metrics.attack_comparator import AttackComparator # noqa from captum.robust._core.metrics.min_param_perturbation import ( # noqa MinParamPerturbation, ) from captum.robust._core.perturbation import Perturbation # n...
#!/usr/bin/env python3 from typing import Any, Callable, Optional, Tuple, Union import torch from captum._utils.common import ( _format_additional_forward_args, _format_output, _format_tensor_into_tuples, _is_tuple, _select_targets, ) from captum._utils.gradient import ( apply_gradient_requirem...
#!/usr/bin/env python3 from typing import Any, Callable, Optional, Tuple, Union import torch import torch.nn.functional as F from captum._utils.common import _format_output, _format_tensor_into_tuples, _is_tuple from captum._utils.typing import TensorOrTupleOfTensorsGeneric from captum.log import log_usage from captum...
#!/usr/bin/env python3 from typing import Callable class Perturbation: r""" All perturbation and attack algorithms extend this class. It enforces its child classes to extend and override core `perturb` method. """ perturb: Callable r""" This method computes and returns the perturbed input...
#!/usr/bin/env python3 import warnings from collections import namedtuple from typing import ( Any, Callable, cast, Dict, Generic, List, NamedTuple, Optional, Tuple, TypeVar, Union, ) from captum._utils.common import ( _expand_additional_forward_args, _format_additio...
#!/usr/bin/env python3 import math from enum import Enum from typing import Any, Callable, cast, Dict, Generator, List, Optional, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _format_additional_forward_args, _reduce_list, ) from captum._utils.typing import T...
#!/usr/bin/env python3 from captum.influence._core.influence import DataInfluence # noqa from captum.influence._core.similarity_influence import SimilarityInfluence # noqa from captum.influence._core.tracincp import TracInCP, TracInCPBase # noqa from captum.influence._core.tracincp_fast_rand_proj import ( TracI...
from abc import ABC, abstractmethod from typing import Tuple import torch from torch import Tensor class NearestNeighbors(ABC): r""" An abstract class to define a nearest neighbors data structure. Classes implementing this interface are intended for computing proponents / opponents in certain impleme...
#!/usr/bin/env python3 import warnings from typing import Any, Callable, List, Optional, Tuple, TYPE_CHECKING, Union import torch import torch.nn as nn from captum._utils.common import _parse_version from captum._utils.progress import progress if TYPE_CHECKING: from captum.influence._core.tracincp import TracInCP...
#!/usr/bin/env python3 import glob import warnings from abc import abstractmethod from os.path import join from typing import ( Any, Callable, Iterator, List, NamedTuple, Optional, Tuple, Type, Union, ) import torch from captum._utils.av import AV from captum._utils.common import _...
#!/usr/bin/env python3 import warnings from functools import partial from typing import Any, Callable, Dict, List, Optional, Tuple, Union import captum._utils.common as common import torch from captum._utils.av import AV from captum.attr import LayerActivation from captum.influence._core.influence import DataInfluenc...
#!/usr/bin/env python3 from abc import ABC, abstractmethod from typing import Any from torch.nn import Module from torch.utils.data import Dataset class DataInfluence(ABC): r""" An abstract class to define model data influence skeleton. """ def __init_(self, model: Module, train_dataset: Dataset, *...
#!/usr/bin/env python3 import threading import warnings from collections import defaultdict from typing import Any, Callable, cast, Dict, Iterator, List, Optional, Tuple, Union import torch from captum._utils.common import _get_module_from_name, _sort_key_list from captum._utils.gradient import _gather_distributed_te...
#!/usr/bin/env python3 from abc import ABC, abstractmethod from typing import Optional, Tuple import torch from torch import Tensor from torch.nn import Module class StochasticGatesBase(Module, ABC): """ Abstract module for Stochastic Gates. Stochastic Gates is a practical solution to add L0 norm regula...
#!/usr/bin/env python3 import math from typing import Optional import torch from captum.module.stochastic_gates_base import StochasticGatesBase from torch import nn, Tensor class GaussianStochasticGates(StochasticGatesBase): """ Stochastic Gates with Gaussian distribution. Stochastic Gates is a practica...
from captum.module.binary_concrete_stochastic_gates import ( # noqa BinaryConcreteStochasticGates, ) from captum.module.gaussian_stochastic_gates import GaussianStochasticGates # noqa from captum.module.stochastic_gates_base import StochasticGatesBase # noqa
#!/usr/bin/env python3 import math from typing import Optional import torch from captum.module.stochastic_gates_base import StochasticGatesBase from torch import nn, Tensor def _torch_empty(batch_size: int, n_gates: int, device: torch.device) -> Tensor: return torch.empty(batch_size, n_gates, device=device) # ...
#!/usr/bin/env python3 from captum.attr._core.dataloader_attr import DataLoaderAttribution # noqa from captum.attr._core.deep_lift import DeepLift, DeepLiftShap # noqa from captum.attr._core.feature_ablation import FeatureAblation # noqa from captum.attr._core.feature_permutation import FeaturePermutation # noqa fr...
#!/usr/bin/env python3 import typing import warnings from typing import Any, Callable, Iterator, Tuple, Union import torch from captum._utils.common import ( _format_additional_forward_args, _format_output, _format_tensor_into_tuples, _reduce_list, ) from captum._utils.typing import ( TargetType, ...
#!/usr/bin/env python3 import torch.nn as nn class Addition_Module(nn.Module): """Custom addition module that uses multiple inputs to assure correct relevance propagation. Any addition in a forward function needs to be replaced with the module before using LRP.""" def __init__(self) -> None: ...
#!/usr/bin/env python3 from enum import Enum from typing import Callable, List, Tuple import torch class Riemann(Enum): left = 1 right = 2 middle = 3 trapezoid = 4 SUPPORTED_RIEMANN_METHODS = [ "riemann_left", "riemann_right", "riemann_middle", "riemann_trapezoid", ] SUPPORTED_METH...
#!/usr/bin/env python3 import inspect from typing import Any import torch.nn as nn class InputIdentity(nn.Module): def __init__(self, input_name: str) -> None: r""" The identity operation Args: input_name (str) The name of the input this layer is associated t...
#!/usr/bin/env python3 import warnings from enum import Enum from typing import Any, Iterable, List, Optional, Tuple, Union import numpy as np from matplotlib import cm, colors, pyplot as plt from matplotlib.collections import LineCollection from matplotlib.colors import LinearSegmentedColormap from matplotlib.figure ...
#!/usr/bin/env python3 from typing import Any, Callable, cast, Generic, List, Tuple, Type, Union import torch import torch.nn.functional as F from captum._utils.common import ( _format_additional_forward_args, _format_tensor_into_tuples, _run_forward, _validate_target, ) from captum._utils.gradient imp...
#!/usr/bin/env python3 import typing from inspect import signature from typing import Any, Callable, List, Tuple, TYPE_CHECKING, Union import torch from captum._utils.common import ( _format_baseline, _format_output, _format_tensor_into_tuples, _validate_input as _validate_input_basic, ) from captum._u...
#!/usr/bin/env python3 from collections import defaultdict from typing import Any, Dict, List, Optional, Union from captum._utils.common import _format_tensor_into_tuples from captum._utils.typing import TargetType, TensorOrTupleOfTensorsGeneric from captum.attr._utils.stat import Stat from captum.attr._utils.summariz...
#!/usr/bin/env python3 from typing import Any, Callable, List, Optional, TYPE_CHECKING import torch from torch import Tensor if TYPE_CHECKING: from captum.attr._utils.summarizer import SummarizerSingleTensor class Stat: """ The Stat class represents a statistic that can be updated and retrieved at a...
#!/usr/bin/env python3 from typing import Dict, List, Optional, Tuple, Type, Union import torch from captum.attr._utils.stat import Count, Max, Mean, Min, MSE, Stat, StdDev, Sum, Var from captum.log import log_usage from torch import Tensor class Summarizer: r""" This class simply wraps over a given a set o...
#!/usr/bin/env python3 from abc import ABC, abstractmethod import torch from ..._utils.common import _format_tensor_into_tuples class PropagationRule(ABC): """ Base class for all propagation rule classes, also called Z-Rule. STABILITY_FACTOR is used to assure that no zero divison occurs. """ S...
#!/usr/bin/env python3 import math from typing import Any, Callable, cast, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_feature_mask, _format_output, _is_tuple, _run_forward, ) from ca...
#!/usr/bin/env python3 from typing import Any, Callable, Tuple, Union import torch from captum._utils.typing import TargetType, TensorOrTupleOfTensorsGeneric from captum.attr._core.feature_ablation import FeatureAblation from captum.log import log_usage from torch import Tensor def _permute_feature(x: Tensor, featur...
#!/usr/bin/env python3 from typing import Any, Callable, Tuple, Union import numpy as np import torch from captum._utils.common import _format_tensor_into_tuples from captum._utils.typing import BaselineType, TargetType, TensorOrTupleOfTensorsGeneric from captum.attr._core.feature_ablation import FeatureAblation from ...
#!/usr/bin/env python3 import typing from typing import Any, Callable, Tuple, Union import numpy as np import torch from captum._utils.common import _is_tuple from captum._utils.typing import ( BaselineType, Literal, TargetType, Tensor, TensorOrTupleOfTensorsGeneric, ) from captum.attr._core.noise_...
#!/usr/bin/env python3 import inspect import math import typing import warnings from typing import Any, Callable, cast, List, Optional, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _flatten_tensor_or_tuple, _format_output, _format_ten...
#!/usr/bin/env python3 from enum import Enum from typing import Any, cast, List, Tuple, Union import torch from captum._utils.common import ( _expand_and_update_additional_forward_args, _expand_and_update_baselines, _expand_and_update_feature_mask, _expand_and_update_target, _format_output, _fo...
#!/usr/bin/env python3 from collections import defaultdict from copy import copy from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union import torch from captum._utils.common import ( _format_baseline, _format_feature_mask, _format_output, _format_tensor_into_tuples, _get_ma...
#!/usr/bin/env python3 import warnings from typing import Any, List, Tuple, Union import torch import torch.nn.functional as F from captum._utils.common import ( _format_output, _format_tensor_into_tuples, _is_tuple, _register_backward_hook, ) from captum._utils.gradient import ( apply_gradient_req...
#!/usr/bin/env python3 import typing from collections import defaultdict from typing import Any, cast, List, Tuple, Union import torch.nn as nn from captum._utils.common import ( _format_output, _format_tensor_into_tuples, _is_tuple, _register_backward_hook, _run_forward, ) from captum._utils.grad...
#!/usr/bin/env python3 import warnings from typing import Any, List, Union import torch from captum._utils.common import _format_output, _format_tensor_into_tuples, _is_tuple from captum._utils.typing import TargetType, TensorOrTupleOfTensorsGeneric from captum.attr._core.guided_backprop_deconvnet import GuidedBackpro...
#!/usr/bin/env python3 import itertools import math import warnings from typing import Any, Callable, Iterable, Sequence, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_feature_mask, _format_out...
#!/usr/bin/env python3 import typing import warnings from typing import Any, Callable, cast, List, Tuple, Union import torch import torch.nn as nn import torch.nn.functional as F from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format...
#!/usr/bin/env python3 from typing import Any, Callable import torch from captum._utils.common import _format_output, _format_tensor_into_tuples, _is_tuple from captum._utils.gradient import ( apply_gradient_requirements, undo_gradient_requirements, ) from captum._utils.typing import TargetType, TensorOrTuple...
#!/usr/bin/env python3 from typing import Any, Callable, Generator, Tuple, Union import torch from captum._utils.models.linear_model import SkLearnLinearRegression from captum._utils.typing import BaselineType, TargetType, TensorOrTupleOfTensorsGeneric from captum.attr._core.lime import construct_feature_mask, Lime f...
#!/usr/bin/env python3 from typing import Any, Callable from captum._utils.common import _format_output, _format_tensor_into_tuples, _is_tuple from captum._utils.gradient import ( apply_gradient_requirements, undo_gradient_requirements, ) from captum._utils.typing import TargetType, TensorOrTupleOfTensorsGener...
#!/usr/bin/env python3 import typing from typing import Any, Callable, List, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_output, _is_tuple, ) from captum._utils.typing import ( BaselineTyp...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union from captum._utils.common import ( _format_additional_forward_args, _format_output, _format_tensor_into_tuples, _is_tuple, ) from captum._utils.gradient import ( _forward_layer_eval_with_neuron_grads, apply_gradient_req...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union from captum._utils.gradient import construct_neuron_grad_fn from captum._utils.typing import TensorOrTupleOfTensorsGeneric from captum.attr._core.guided_backprop_deconvnet import Deconvolution, GuidedBackprop from captum.attr._utils.attributio...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union import torch from captum._utils.common import _verify_select_neuron from captum._utils.gradient import _forward_layer_eval from captum._utils.typing import BaselineType, TensorOrTupleOfTensorsGeneric from captum.attr._core.feature_ablation imp...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union from captum._utils.gradient import construct_neuron_grad_fn from captum._utils.typing import TensorOrTupleOfTensorsGeneric from captum.attr._core.gradient_shap import GradientShap from captum.attr._utils.attribution import GradientAttribution,...
#!/usr/bin/env python3 from typing import Any, Callable, cast, Tuple, Union from captum._utils.gradient import construct_neuron_grad_fn from captum._utils.typing import BaselineType, TensorOrTupleOfTensorsGeneric from captum.attr._core.deep_lift import DeepLift, DeepLiftShap from captum.attr._utils.attribution import ...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union from captum._utils.gradient import construct_neuron_grad_fn from captum._utils.typing import TensorOrTupleOfTensorsGeneric from captum.attr._core.integrated_gradients import IntegratedGradients from captum.attr._utils.attribution import Gradie...
#!/usr/bin/env python3 import warnings from typing import Any, Callable, List, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_output, _is_tuple, _verify_select_neuron, ) from captum._utils.gr...
#!/usr/bin/env python3 from typing import Any, Callable, List, Tuple, Union import torch from captum._utils.common import ( _expand_additional_forward_args, _expand_target, _format_additional_forward_args, _format_output, ) from captum._utils.gradient import compute_layer_gradients_and_eval from captum...
#!/usr/bin/env python3 import typing from typing import Any, cast, List, Tuple, Union from captum._utils.common import ( _format_tensor_into_tuples, _reduce_list, _sort_key_list, ) from captum._utils.gradient import ( apply_gradient_requirements, compute_gradients, undo_gradient_requirements, )...