input stringlengths 11 7.65k | target stringlengths 22 8.26k |
|---|---|
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, user_names_str):
if not hasattr(context, 'users'):
context.users = {}
user_names = [name.strip() for name in re.split('and|,', user_names_str)]
for user_name in user_names:
token = 'fake_token_' + user_name
user_id = context.helpers.create_test_user(user_name,... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def fixed_padding(inputs, kernel_size, data_format):
"""Pads the input along the spatial dimensions independently of input size.
Args:
inputs: A tensor of size [batch, channels, height_in, width_in] or [batch,
height_in, width_in, channels] depending on data_format.
kernel_size: The kernel to be used... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, user_name):
context.token = context.users[user_name]['token'] |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def conv2d_fixed_padding(inputs, filters, kernel_size, strides, data_format,
name):
"""Strided 2-D convolution with explicit padding."""
# The padding is consistent and is based only on `kernel_size`, not on the
# dimensions of `inputs` (as opposed to using `tf.layers.conv2d` alone).
if... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, user_name, count):
context.helpers.load_postcards(user_name, count) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _building_block_v2(inputs, filters, training, projection_shortcut, strides,
data_format, name):
"""A single block for ResNet v2, without a bottleneck.
Batch normalization then ReLu then convolution as described by:
Identity Mappings in Deep Residual Networks
https://arxiv.org/pdf... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, rel_url):
context.request = LazyRequest(
'GET', context.helpers.url(rel_url), context.token) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def projection_shortcut(inputs, name):
return conv2d_fixed_padding(
inputs=inputs,
filters=filters_out,
kernel_size=1,
strides=strides,
data_format=data_format,
name=name) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, rel_url):
context.request = LazyRequest(
'POST', context.helpers.url(rel_url), context.token) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self,
resnet_size,
bottleneck,
num_classes,
num_filters,
kernel_size,
conv_stride,
first_pool_size,
first_pool_stride,
block_sizes,
block_strides,
... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, name, field):
context.request.add_file(context.helpers.file_path(name), field) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _custom_dtype_getter(self, # pylint: disable=keyword-arg-before-vararg
getter,
name,
shape=None,
dtype=DEFAULT_DTYPE,
*args,
**kwargs):
"""Creates va... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context):
data = json.loads(context.text)
receiver_name = re.match(r"\<(\w+)'s id\>", data['receiver']).group(1)
data['receiver'] = context.users[receiver_name]['id']
context.request.add_data(data) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _model_variable_scope(self):
"""Returns a variable scope that the model should be created under.
If self.dtype is a castable type, model variable will be created in fp32
then cast to self.dtype before being used.
Returns:
A variable scope for the model.
"""
return tf.compat.v1.varia... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, state, code):
context.response = context.request.send()
context.response.status_code.should.equal(int(code)) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __call__(self, inputs, training):
"""Add operations to classify a batch of input images.
Args:
inputs: A Tensor representing a batch of input images.
training: A boolean. Set to True to add operations required only when
training the classifier.
Returns:
A logits Tensor with s... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def step_impl(context, count):
cnt = len(context.response.json())
cnt.should.equal(int(count)) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def noun_chunks(doclike: Union[Doc, Span]) -> Iterator[Span]:
"""
Detect base noun phrases from a dependency parse. Works on both Doc and Span.
"""
# fmt: off
labels = ["nsubj", "nsubj:pass", "obj", "iobj", "ROOT", "appos", "nmod", "nmod:poss"]
# fmt: on
doc = doclike.doc # Ensure works on ... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, leaf):
self.leaf = leaf
self.lchild = None
self.rchild = None |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def get_leafs(self):
if self.lchild == None and self.rchild == None:
return [self.leaf]
else:
return self.lchild.get_leafs()+self.rchild.get_leafs() |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def get_level(self, level, queue):
if queue == None:
queue = []
if level == 1:
queue.push(self)
else:
if self.lchild != None:
self.lchild.get_level(level-1, queue)
if self.rchild != None:
self.rchild.get_level(level-1, queue)
return queue |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def paint(self, c):
self.leaf.paint(c)
if self.lchild != None:
self.lchild.paint(c)
if self.rchild != None:
self.rchild.paint(c) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, x, y, w, h):
self.x = x
self.y = y
self.w = w
self.h = h
self.center = (self.x+int(self.w/2),self.y+int(self.h/2))
self.distance_from_center = sqrt((self.center[0]-MAP_WIDTH/2)**2 + (self.center[1]-MAP_HEIGHT/2)**2) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def paint(self, c):
c.stroke_rectangle(self.x, self.y, self.w, self.h) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def draw_path(self,c,container):
c.path(self.center[0],self.center[1],container.center[0],container.center[1]) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, w, h, color = "empty"):
self.board = zeros((h,w), dtype=uint8)
self.w = w
self.h = h
self.set_brush(color) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def set_brush(self, code):
self.color = self.brushes[code] |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def stroke_rectangle(self, x, y, w, h):
self.line(x,y,w,True)
self.line(x,y+h-1,w,True)
self.line(x,y,h,False)
self.line(x+w-1,y,h,False) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def filled_rectangle(self, x, y, w, h):
self.board[y:y+h,x:x+w] = self.color |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def line(self, x, y, length, horizontal):
if horizontal:
self.board[y,x:x+length] = self.color
else:
self.board[y:y+length,x] = self.color |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def path(self,x1,y1,x2,y2):
self.board[y1:y2+1,x1:x2+1] = self.color |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def circle(self,x,y,r):
for x_offset in range(-r,r+1):
for y_offset in range(-r,r+1):
if sqrt(x_offset**2+y_offset**2)<r:
self.board[x+x_offset,y+y_offset] = self.color |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def draw(self):
im = Image.fromarray(self.board)
im.save(MAP_NAME) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __str__(self):
return str(self.board) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, container):
self.x = container.x+randint(1, floor(container.w/3))
self.y = container.y+randint(1, floor(container.h/3))
self.w = container.w-(self.x-container.x)
self.h = container.h-(self.y-container.y)
self.w -= randint(0,floor(self.w/3))
self.h -= randint(0,floor(self.w/3))
self.envi... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def paint(self,c):
c.filled_rectangle(self.x, self.y,self.w, self.h) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _split_vertical(container):
r1 = None
r2 = None
min_w = int(W_RATIO*container.h)+1
if container.w < 2*min_w:
return None
r1 = Container(container.x,container.y,randint(min_w, container.w-min_w),container.h)
r2 = Container(container.x+r1.w,container.y,container.w-r1.w,container.h)
return [r1, r2] |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _split_horizontal(container):
r1 = None
r2 = None
min_h = int(H_RATIO*container.w)+1
if container.h < 2*min_h:
return None
r1 = Container(container.x,container.y,container.w,randint(min_h, container.h-min_h))
r2 = Container(container.x,container.y+r1.h,container.w,container.h-r1.h)
return [r1, r2] |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def split_container(container, iter):
root = Tree(container)
if iter != 0:
sr = random_split(container)
if sr!=None:
root.lchild = split_container(sr[0], iter-1)
root.rchild = split_container(sr[1], iter-1)
return root |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def draw_paths(c, tree):
if tree.lchild == None or tree.rchild == None:
return
tree.lchild.leaf.draw_path(c, tree.rchild.leaf)
draw_paths(c, tree.lchild)
draw_paths(c, tree.rchild) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def init(num_players):
global MAP_WIDTH,MAP_HEIGHT,N_ITERATIONS,H_RATIO,W_RATIO,MIN_ROOM_SIDE,CENTER_HUB_HOLE,CENTER_HUB_RADIO,MAP_NAME
MAP_WIDTH=int(500*sqrt(num_players))
MAP_HEIGHT=MAP_WIDTH
N_ITERATIONS=log(MAP_WIDTH*100,2)
H_RATIO=0.49
W_RATIO=H_RATIO
MIN_ROOM_SIDE = 32
CENTER_HUB_HOLE = 32
CENTER_HUB_RAD... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def get_attn(attn_type):
if isinstance(attn_type, torch.nn.Module):
return attn_type
module_cls = None
if attn_type is not None:
if isinstance(attn_type, str):
attn_type = attn_type.lower()
# Lightweight attention modules (channel and/or coarse spatial).
#... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def main():
argument_spec = ec2_argument_spec()
argument_spec.update(dict(
region = dict(required=True,
aliases = ['aws_region', 'ec2_region']),
owner = dict(required=False, default=None),
ami_id = dict(required=False),
ami_tags = dict(required=Fal... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def start(self, action_name: str) -> None:
"""Defines how to start recording an action.""" |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def install_secret_key(app, filename='secret_key'):
"""Configure the SECRET_KEY from a file
in the instance directory.
If the file does not exist, print instructions
to create it from a shell with a random key,
then exit.
"""
filename = os.path.join(app.instance_path, filename)
try:
... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def stop(self, action_name: str) -> None:
"""Defines how to record the duration once an action is complete.""" |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def not_found(error):
return render_template('404.html'), 404 |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def summary(self) -> str:
"""Create profiler summary in text format.""" |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def after_request(response):
response.headers.add('X-Test', 'This is only test.')
response.headers.add('Access-Control-Allow-Origin', '*') # TODO: set to real origin
return response |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def setup(self, **kwargs: Any) -> None:
"""Execute arbitrary pre-profiling set-up steps as defined by subclass.""" |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def teardown(self, **kwargs: Any) -> None:
"""Execute arbitrary post-profiling tear-down steps as defined by subclass.""" |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(
self,
dirpath: Optional[Union[str, Path]] = None,
filename: Optional[str] = None,
) -> None:
self.dirpath = dirpath
self.filename = filename
self._output_file: Optional[TextIO] = None
self._write_stream: Optional[Callable] = None
self._l... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def profile(self, action_name: str) -> Generator:
"""
Yields a context manager to encapsulate the scope of a profiled action.
Example::
with self.profile('load training data'):
# load training data code
The profiler will start once you've entered the contex... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def profile_iterable(self, iterable: Iterable, action_name: str) -> Generator:
iterator = iter(iterable)
while True:
try:
self.start(action_name)
value = next(iterator)
self.stop(action_name)
yield value
except StopI... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _rank_zero_info(self, *args, **kwargs) -> None:
if self._local_rank in (None, 0):
log.info(*args, **kwargs) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _prepare_filename(
self, action_name: Optional[str] = None, extension: str = ".txt", split_token: str = "-"
) -> str:
args = []
if self._stage is not None:
args.append(self._stage)
if self.filename:
args.append(self.filename)
if self._local_rank is... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _prepare_streams(self) -> None:
if self._write_stream is not None:
return
if self.filename:
filepath = os.path.join(self.dirpath, self._prepare_filename())
fs = get_filesystem(filepath)
file = fs.open(filepath, "a")
self._output_file = file... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def describe(self) -> None:
"""Logs a profile report after the conclusion of run."""
# there are pickling issues with open file handles in Python 3.6
# so to avoid them, we open and close the files within this function
# by calling `_prepare_streams` and `teardown`
self._prepare_... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _stats_to_str(self, stats: Dict[str, str]) -> str:
stage = f"{self._stage.upper()} " if self._stage is not None else ""
output = [stage + "Profiler Report"]
for action, value in stats.items():
header = f"Profile stats for: {action}"
if self._local_rank is not None:
... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def setup(
self, stage: Optional[str] = None, local_rank: Optional[int] = None, log_dir: Optional[str] = None
) -> None:
"""Execute arbitrary pre-profiling set-up steps."""
self._stage = stage
self._local_rank = local_rank
self._log_dir = log_dir
self.dirpath = self.d... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def teardown(self, stage: Optional[str] = None) -> None:
"""
Execute arbitrary post-profiling tear-down steps.
Closes the currently open file and stream.
"""
self._write_stream = None
if self._output_file is not None:
self._output_file.close()
sel... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __del__(self) -> None:
self.teardown(stage=self._stage) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def start(self, action_name: str) -> None:
raise NotImplementedError |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def stop(self, action_name: str) -> None:
raise NotImplementedError |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def summary(self) -> str:
raise NotImplementedError |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def local_rank(self) -> int:
return 0 if self._local_rank is None else self._local_rank |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def start(self, action_name: str) -> None:
pass |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def stop(self, action_name: str) -> None:
pass |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, model, data):
# try and import pytorch
global torch
if torch is None:
import torch
if version.parse(torch.__version__) < version.parse("0.4"):
warnings.warn("Your PyTorch version is older than 0.4 and not supported.")
# check if... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def __init__(self, reddit, term, config, oauth, url=None, submission=None):
super(SubmissionPage, self).__init__(reddit, term, config, oauth)
self.controller = SubmissionController(self, keymap=config.keymap)
if url:
self.content = SubmissionContent.from_url(
reddit... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def add_target_handle(self, layer):
input_handle = layer.register_forward_hook(get_target_input)
self.target_handle = input_handle |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def toggle_comment(self):
"Toggle the selected comment tree between visible and hidden"
current_index = self.nav.absolute_index
self.content.toggle(current_index)
# This logic handles a display edge case after a comment toggle. We
# want to make sure that when we re-draw the pa... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def add_handles(self, model, forward_handle, backward_handle):
"""
Add handles to all non-container layers in the model.
Recursively for non-container layers
"""
handles_list = []
model_children = list(model.children())
if model_children:
for child in ... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def exit_submission(self):
"Close the submission and return to the subreddit page"
self.active = False |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def remove_attributes(self, model):
"""
Removes the x and y attributes which were added by the forward handles
Recursively searches for non-container layers
"""
for child in model.children():
if 'nn.modules.container' in str(type(child)):
self.remove_a... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def refresh_content(self, order=None, name=None):
"Re-download comments and reset the page index"
order = order or self.content.order
url = name or self.content.name
with self.term.loader('Refreshing page'):
self.content = SubmissionContent.from_url(
self.re... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def gradient(self, idx, inputs):
self.model.zero_grad()
X = [x.requires_grad_() for x in inputs]
outputs = self.model(*X)
selected = [val for val in outputs[:, idx]]
grads = []
if self.interim:
interim_inputs = self.layer.target_input
for idx, inpu... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def prompt_subreddit(self):
"Open a prompt to navigate to a different subreddit"
name = self.term.prompt_input('Enter page: /')
if name is not None:
with self.term.loader('Loading page'):
content = SubredditContent.from_name(
self.reddit, name, se... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def shap_values(self, X, ranked_outputs=None, output_rank_order="max", check_additivity=False):
# X ~ self.model_input
# X_data ~ self.data
# check if we have multiple inputs
if not self.multi_input:
assert type(X) != list, "Expected a single tensor model input!"
... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def open_link(self):
"Open the selected item with the webbrowser"
data = self.get_selected_item()
url = data.get('permalink')
if url:
self.term.open_browser(url)
else:
self.term.flash() |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def deeplift_grad(module, grad_input, grad_output):
"""The backward hook which computes the deeplift
gradient for an nn.Module
"""
# first, get the module type
module_type = module.__class__.__name__
# first, check the module is supported
if module_type in op_handler:
if op_handler[m... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def open_pager(self):
"Open the selected item with the system's pager"
data = self.get_selected_item()
if data['type'] == 'Submission':
text = '\n\n'.join((data['permalink'], data['text']))
self.term.open_pager(text)
elif data['type'] == 'Comment':
tex... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def add_interim_values(module, input, output):
"""The forward hook used to save interim tensors, detached
from the graph. Used to calculate the multipliers
"""
try:
del module.x
except AttributeError:
pass
try:
del module.y
except AttributeError:
pass
modu... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def add_comment(self):
"""
Submit a reply to the selected item.
Selected item:
Submission - add a top level comment
Comment - add a comment reply
"""
data = self.get_selected_item()
if data['type'] == 'Submission':
body = data['text']... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def get_target_input(module, input, output):
"""A forward hook which saves the tensor - attached to its graph.
Used if we want to explain the interim outputs of a model
"""
try:
del module.target_input
except AttributeError:
pass
setattr(module, 'target_input', input) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def delete_comment(self):
"Delete the selected comment"
if self.get_selected_item()['type'] == 'Comment':
self.delete_item()
else:
self.term.flash() |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def deeplift_tensor_grad(grad):
return_grad = complex_module_gradients[-1]
del complex_module_gradients[-1]
return return_grad |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def comment_urlview(self):
data = self.get_selected_item()
comment = data.get('body') or data.get('text') or data.get('url_full')
if comment:
self.term.open_urlview(comment)
else:
self.term.flash() |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def passthrough(module, grad_input, grad_output):
"""No change made to gradients"""
return None |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _draw_item(self, win, data, inverted):
if data['type'] == 'MoreComments':
return self._draw_more_comments(win, data)
elif data['type'] == 'HiddenComment':
return self._draw_more_comments(win, data)
elif data['type'] == 'Comment':
return self._draw_comment... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def maxpool(module, grad_input, grad_output):
pool_to_unpool = {
'MaxPool1d': torch.nn.functional.max_unpool1d,
'MaxPool2d': torch.nn.functional.max_unpool2d,
'MaxPool3d': torch.nn.functional.max_unpool3d
}
pool_to_function = {
'MaxPool1d': torch.nn.functional.max_pool1d,
... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _draw_comment(self, win, data, inverted):
n_rows, n_cols = win.getmaxyx()
n_cols -= 1
# Handle the case where the window is not large enough to fit the text.
valid_rows = range(0, n_rows)
offset = 0 if not inverted else -(data['n_rows'] - n_rows)
# If there isn't e... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def linear_1d(module, grad_input, grad_output):
"""No change made to gradients."""
return None |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _draw_more_comments(self, win, data):
n_rows, n_cols = win.getmaxyx()
n_cols -= 1
self.term.add_line(win, '{body}'.format(**data), 0, 1)
self.term.add_line(
win, ' [{count}]'.format(**data), attr=curses.A_BOLD)
attr = Color.get_level(data['level'])
self... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def nonlinear_1d(module, grad_input, grad_output):
delta_out = module.y[: int(module.y.shape[0] / 2)] - module.y[int(module.y.shape[0] / 2):]
delta_in = module.x[: int(module.x.shape[0] / 2)] - module.x[int(module.x.shape[0] / 2):]
dup0 = [2] + [1 for i in delta_in.shape[1:]]
# handles numerical instab... |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def wait_time_gen():
count = 0
while True:
rand = random.randrange(round(interval.total_seconds()))
tmp = round(start + interval.total_seconds() * count + rand - loop.time())
yield tmp
count += 1 |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def every_day(job, loop=None):
return every(job, timedelta=timedelta(days=1), loop=loop) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def every_week(job, loop=None):
return every(job, timedelta=timedelta(days=7), loop=loop) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _nearest_weekday(weekday):
return datetime.now() + timedelta(days=(weekday - datetime.now().weekday()) % 7) |
def save(self, doc):
if doc.id:
return self.update(doc)
else:
return self.insert(doc) | def _every_weekday(job, weekday, loop=None):
return every(job, timedelta=timedelta(days=7), start_at=_nearest_weekday(weekday), loop=loop) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.