entry_point
stringlengths
1
65
original_triton_code
stringlengths
4.5k
619k
python_code
stringlengths
208
60.9k
triton_code
stringlengths
1.15k
275k
repo_name
stringlengths
7
115
module_name
stringlengths
1
65
synthetic
bool
1 class
uuid
int64
0
18.5k
licenses
listlengths
1
6
stars
int64
0
19.8k
sha
stringlengths
40
40
repo_link
stringlengths
72
180
pytorch_code
stringlengths
200
4.05k
Features_2_to_1
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.optim import torch.nn as nn class Features_2_to_1(nn.Module): def __init__(self): """ take a batch (bs, n_vertices, n_vertices, in_features) and returns (bs, n_vertices, basis * in_features) where basis = 5 """ super().__init__() def ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.optim import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.ass...
MauTrib/gnn-en-folie
Features_2_to_1
false
831
[ "Apache-2.0" ]
0
3ca639919a2b285a41641717f4131107c015b510
https://github.com/MauTrib/gnn-en-folie/tree/3ca639919a2b285a41641717f4131107c015b510
import torch import torch.optim import torch.nn as nn class Model(nn.Module): def __init__(self): """ take a batch (bs, n_vertices, n_vertices, in_features) and returns (bs, n_vertices, basis * in_features) where basis = 5 """ super().__init__() def forward(se...
Scale
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class Scale(nn.Module): def __init__(self, scale=30): super(Scale, self).__init__() self.scale = scale def forward(self, x): return x * self.scale def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], {}]...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
MickeyZeng/Data-Visualization
Scale
false
832
[ "MIT" ]
0
c7005d1096545d7a5eb96dd0c9bc13e874d42fa4
https://github.com/MickeyZeng/Data-Visualization/tree/c7005d1096545d7a5eb96dd0c9bc13e874d42fa4
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, scale=30): super().__init__() self.scale = scale def forward(self, x): return x * self.scale def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []
MutualBiAffineAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from torch.autograd import * class MutualBiAffineAttention(nn.Module): """ Mutual BiAffine Attention between 2 kinds of features. """ def __init__(self, hidden_size): super(MutualBiAffineAttention, self).__init__() self.linear1 = nn.Linear(2 * hidden...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math im...
Maxi-0902/DRAN
MutualBiAffineAttention
false
833
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
import torch import torch.nn as nn from torch.autograd import * class Model(nn.Module): """ Mutual BiAffine Attention between 2 kinds of features. """ def __init__(self, hidden_size): super().__init__() self.linear1 = nn.Linear(2 * hidden_size, hidden_size) self.linear2 = nn.L...
DownBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ConvBlock(nn.Module): def __init__(self, input_size, output_size, kernel_size=3, stride=1, padding=1, bias=True, norm=None): super(ConvBlock, self).__init__() self.conv = nn.Conv2d(input_size, output_size, kernel_size, stride, padding, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
MatusBako/MakeFacesGreatAgain
DownBlock
false
834
[ "MIT" ]
0
e4941a8460db79dec566ed02d4b23eafb416a6db
https://github.com/MatusBako/MakeFacesGreatAgain/tree/e4941a8460db79dec566ed02d4b23eafb416a6db
import torch import torch.nn as nn class ConvBlock(nn.Module): def __init__(self, input_size, output_size, kernel_size=3, stride=1, padding=1, bias=True, norm=None): super().__init__() self.conv = nn.Conv2d(input_size, output_size, kernel_size, stride, padding, bias=bias) ...
RewardCriterion
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torch.autograd import * def to_contiguous(tensor): if tensor.is_contiguous(): return tensor else: return tensor.contiguous() class RewardCriterion(nn.Module): def __init__(self): super(RewardCriterion, self).__init__() def forward(sel...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from torch.autograd import * assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torc...
Maxi-0902/DRAN
RewardCriterion
false
835
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
import torch import torch.nn as nn from torch.autograd import * def to_contiguous(tensor): if tensor.is_contiguous(): return tensor else: return tensor.contiguous() class Model(nn.Module): def __init__(self): super().__init__() def forward(self, input, seq, reward, gpn_loss...
SelfAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn as nn import torch.utils.data.distributed class SelfAttention(nn.Module): """ Self SelfAttention Layer Given $X\\in \\mathbb{R}^{n imes in_feature}$, the attention is calculated by: $a=Softmax(W_2tanh(W_1X))$, where $W_1 \\in \\mathbb{R}^{hidden imes in_fea...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
MatthewMasters/grover
SelfAttention
false
836
[ "MIT" ]
0
737a340754bc4c63134ef84019a0a84023fd69a3
https://github.com/MatthewMasters/grover/tree/737a340754bc4c63134ef84019a0a84023fd69a3
import torch from torch import nn as nn import torch.utils.data.distributed class Model(nn.Module): """ Self SelfAttention Layer Given $X\\in \\mathbb{R}^{n imes in_feature}$, the attention is calculated by: $a=Softmax(W_2tanh(W_1X))$, where $W_1 \\in \\mathbb{R}^{hidden imes in_feature}$, ...
DWConv
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class DWConv(nn.Module): def __init__(self, dim=768): super(DWConv, self).__init__() self.dwconv = nn.Conv2d(dim, dim, 3, 1, 1, bias=True, groups=dim) def forward(self, x): x = self.dwconv(x) return x def get_inputs(): return [torch.ra...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
LSH9832/MyPythonModules
DWConv
false
837
[ "MIT" ]
0
442566a0fbd6ebe2bc20b6914686a1e2663d10c0
https://github.com/LSH9832/MyPythonModules/tree/442566a0fbd6ebe2bc20b6914686a1e2663d10c0
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, dim=768): super().__init__() self.dwconv = nn.Conv2d(dim, dim, 3, 1, 1, bias=True, groups=dim) def forward(self, x): x = self.dwconv(x) return x def get_inputs(): return [torch.rand([4, 768, 6...
Squareplus
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch as t import torch.nn as nn class Squareplus(nn.Module): def __init__(self, a=2): super().__init__() self.a = a def forward(self, x): """The 'squareplus' activation function: has very similar properties to softplus, but is far cheaper computationally....
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
MaximeRobeyns/BDRL
Squareplus
false
838
[ "Apache-2.0" ]
0
55e295d5aaca6745d35525114b472ad118c14a6d
https://github.com/MaximeRobeyns/BDRL/tree/55e295d5aaca6745d35525114b472ad118c14a6d
import torch import torch as t import torch.nn as nn class Model(nn.Module): def __init__(self, a=2): super().__init__() self.a = a def forward(self, x): """The 'squareplus' activation function: has very similar properties to softplus, but is far cheaper computationally. ...
DiceBCELoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn from torch.nn import functional as F class DiceBCELoss(nn.Module): def __init__(self, weight=None, size_average=True): super(DiceBCELoss, self).__init__() def forward(self, inputs, targets, smooth=1): inputs = torch.sigmoid(inputs) inputs = inputs.vi...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math from torch ...
Michaelistaken/PathPretrain
DiceBCELoss
false
839
[ "MIT" ]
0
650b7eb02e67f6d864d81808eb7230c48fe6946a
https://github.com/Michaelistaken/PathPretrain/tree/650b7eb02e67f6d864d81808eb7230c48fe6946a
import torch from torch import nn from torch.nn import functional as F class Model(nn.Module): def __init__(self, weight=None, size_average=True): super().__init__() def forward(self, inputs, targets, smooth=1): inputs = torch.sigmoid(inputs) inputs = inputs.view(-1) targets ...
DiceLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class DiceLoss(nn.Module): def __init__(self, weight=None, size_average=True): super(DiceLoss, self).__init__() def forward(self, inputs, targets): intersection = (inputs * targets).sum() dice = (2.0 * intersection + 1e-05) / (inputs.sum() + targets...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
MohannadEhabBarakat/U-2-Net
DiceLoss
false
840
[ "Apache-2.0" ]
0
89a4eba7a565e7afcd4ac04b11b55099ebef687c
https://github.com/MohannadEhabBarakat/U-2-Net/tree/89a4eba7a565e7afcd4ac04b11b55099ebef687c
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, weight=None, size_average=True): super().__init__() def forward(self, inputs, targets): intersection = (inputs * targets).sum() dice = (2.0 * intersection + 1e-05) / (inputs.sum() + targets.sum() + ...
MSELoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import functools import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss ten...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import functools import torch.nn as nn import torch.nn.functional as F assert_size_stride...
Min-Sheng/mmregression
MSELoss
false
841
[ "Apache-2.0" ]
0
6d70383d89ccb3dea7f425b665c2a184d014a99f
https://github.com/Min-Sheng/mmregression/tree/6d70383d89ccb3dea7f425b665c2a184d014a99f
import functools import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss ten...
UpBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ConvBlock(nn.Module): def __init__(self, input_size, output_size, kernel_size=3, stride=1, padding=1, bias=True, norm=None): super(ConvBlock, self).__init__() self.conv = nn.Conv2d(input_size, output_size, kernel_size, stride, padding, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
MatusBako/MakeFacesGreatAgain
UpBlock
false
842
[ "MIT" ]
0
e4941a8460db79dec566ed02d4b23eafb416a6db
https://github.com/MatusBako/MakeFacesGreatAgain/tree/e4941a8460db79dec566ed02d4b23eafb416a6db
import torch import torch.nn as nn class ConvBlock(nn.Module): def __init__(self, input_size, output_size, kernel_size=3, stride=1, padding=1, bias=True, norm=None): super().__init__() self.conv = nn.Conv2d(input_size, output_size, kernel_size, stride, padding, bias=bias) ...
AdaptiveAveragePooling
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class AdaptiveAveragePooling(nn.Module): """Adaptive Pooling neck. Args: dim (int): Dimensions of each sample channel, can be one of {1, 2, 3}. Default: 2 output_size (int | tuple): output size, If dim equals to 1: output_size is a ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Min-Sheng/mmregression
AdaptiveAveragePooling
false
843
[ "Apache-2.0" ]
0
6d70383d89ccb3dea7f425b665c2a184d014a99f
https://github.com/Min-Sheng/mmregression/tree/6d70383d89ccb3dea7f425b665c2a184d014a99f
import torch import torch.nn as nn class Model(nn.Module): """Adaptive Pooling neck. Args: dim (int): Dimensions of each sample channel, can be one of {1, 2, 3}. Default: 2 output_size (int | tuple): output size, If dim equals to 1: output_size is a single integer. ...
DiceLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F def flatten(tensor): """Flattens a given tensor such that the channel axis is first. The shapes are transformed as follows: (N, C, D, H, W) -> (C, N * D * H * W) """ C = tensor.size(1) axis_order = (1, 0) + tuple(range(2, te...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
MikoyChinese/Learn
DiceLoss
false
844
[ "Apache-2.0" ]
0
c482b1e84496279935b5bb2cfc1e6d78e2868c63
https://github.com/MikoyChinese/Learn/tree/c482b1e84496279935b5bb2cfc1e6d78e2868c63
import torch import torch.nn as nn import torch.nn.functional as F def flatten(tensor): """Flattens a given tensor such that the channel axis is first. The shapes are transformed as follows: (N, C, D, H, W) -> (C, N * D * H * W) """ C = tensor.size(1) axis_order = (1, 0) + tuple(range(2, te...
Residual_Block
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from torch import add class Residual_Block(nn.Module): def __init__(self): super(Residual_Block, self).__init__() self.conv1 = nn.Conv2d(in_channels=64, out_channels=64, kernel_size =3, padding=1, bias=False) self.in1 = nn.InstanceNorm2d(64, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
MatusBako/MakeFacesGreatAgain
Residual_Block
false
845
[ "MIT" ]
0
e4941a8460db79dec566ed02d4b23eafb416a6db
https://github.com/MatusBako/MakeFacesGreatAgain/tree/e4941a8460db79dec566ed02d4b23eafb416a6db
import torch import torch.nn as nn from torch import add class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(in_channels=64, out_channels=64, kernel_size =3, padding=1, bias=False) self.in1 = nn.InstanceNorm2d(64, affine=True) self.rel...
ScaledDotProductAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductAttention(nn.Module): """ Scaled dot-product attention """ def __init__(self, d_model, d_k, d_v, h): """ :param d_model: Output dimensionality of the model :param d_k: Dime...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Maxi-0902/DRAN
ScaledDotProductAttention
false
846
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
import torch import numpy as np import torch.nn as nn from torch.autograd import * class Model(nn.Module): """ Scaled dot-product attention """ def __init__(self, d_model, d_k, d_v, h): """ :param d_model: Output dimensionality of the model :param d_k: Dimensionality of querie...
activation_quantize_fn
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch.nn as nn def uniform_quantize(k): class qfn(torch.autograd.Function): @staticmethod def forward(ctx, input): if k == 32: out = input elif k == 1: out = torch.sign(input) else: ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.utils.data import torch.nn as nn assert_size_stride = torch._C._dy...
MohammedHAlali/pytorch_DoReFaNet
activation_quantize_fn
false
847
[ "MIT" ]
0
d208089b9172f02c09cc6633158ed5b5d6cd7f1e
https://github.com/MohammedHAlali/pytorch_DoReFaNet/tree/d208089b9172f02c09cc6633158ed5b5d6cd7f1e
import torch import torch.utils.data import torch.nn as nn def uniform_quantize(k): class qfn(torch.autograd.Function): @staticmethod def forward(ctx, input): if k == 32: out = input elif k == 1: out = torch.sign(input) else: ...
FeedForward
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import n...
Misuzu-Kurenai/mlp-singer
FeedForward
false
848
[ "MIT" ]
0
416451045bb9b3965aaf496e84a8b45332a6ba59
https://github.com/Misuzu-Kurenai/mlp-singer/tree/416451045bb9b3965aaf496e84a8b45332a6ba59
import torch import torch.nn.functional as F from torch import nn class Model(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn.Linea...
ScaledDotProductWithBoxAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductWithBoxAttention(nn.Module): """ Scaled dot-product attention with box """ def __init__(self, d_model, d_k, d_v, h, dropout=0.1, comment=None): """ :param d_model: Output dimension...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Maxi-0902/DRAN
ScaledDotProductWithBoxAttention
false
849
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
import torch import numpy as np import torch.nn as nn from torch.autograd import * class Model(nn.Module): """ Scaled dot-product attention with box """ def __init__(self, d_model, d_k, d_v, h, dropout=0.1, comment=None): """ :param d_model: Output dimensionality of the model ...
GlobalAveragePooling
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class GlobalAveragePooling(nn.Module): def __init__(self): super(GlobalAveragePooling, self).__init__() def forward(self, feat): num_channels = feat.size(1) return F.avg_pool2d(feat, (feat.size(2), feat.size(3))).view...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
MrChenFeng/Project-Template
GlobalAveragePooling
false
850
[ "MIT" ]
0
42a335c6abb710bbae6407cbb0ca461533bc12f9
https://github.com/MrChenFeng/Project-Template/tree/42a335c6abb710bbae6407cbb0ca461533bc12f9
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() def forward(self, feat): num_channels = feat.size(1) return F.avg_pool2d(feat, (feat.size(2), feat.size(3))).view(-1, num_channels) def get_...
RadialPredictionLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class RadialPredictionLayer(torch.nn.Module): """ The RPL classification layer with fixed prototypes """ def __init__(self, in_features, out_features): super(RadialPredictionLayer, self).__init__() self.in_features = in_features self.out_features...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
Monkso/RPL-Softmax_RoadSigns
RadialPredictionLayer
false
851
[ "MIT" ]
0
3df929d779ff02ec796e717659943bb46311ba0f
https://github.com/Monkso/RPL-Softmax_RoadSigns/tree/3df929d779ff02ec796e717659943bb46311ba0f
import torch import torch.nn as nn class Model(torch.nn.Module): """ The RPL classification layer with fixed prototypes """ def __init__(self, in_features, out_features): super().__init__() self.in_features = in_features self.out_features = out_features self.prototypes = n...
BBoxTransform
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class BBoxTransform(nn.Module): def forward(self, anchors, regression): """ Args: anchors: [batch_size, boxes, (y1, x1, y2, x2)] regression: [batch_size, boxes, (dy, dx, dh, dw)] """ y_centers_a = (anchors[..., 0] + anchor...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert...
MikoyChinese/Learn
BBoxTransform
false
852
[ "Apache-2.0" ]
0
c482b1e84496279935b5bb2cfc1e6d78e2868c63
https://github.com/MikoyChinese/Learn/tree/c482b1e84496279935b5bb2cfc1e6d78e2868c63
import torch import torch.nn as nn class Model(nn.Module): def forward(self, anchors, regression): """ Args: anchors: [batch_size, boxes, (y1, x1, y2, x2)] regression: [batch_size, boxes, (dy, dx, dh, dw)] """ y_centers_a = (anchors[..., 0] + anchors[..., 2...
weight_quantize_fn
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch.nn as nn def uniform_quantize(k): class qfn(torch.autograd.Function): @staticmethod def forward(ctx, input): if k == 32: out = input elif k == 1: out = torch.sign(input) else: ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
MohammedHAlali/pytorch_DoReFaNet
weight_quantize_fn
false
853
[ "MIT" ]
0
d208089b9172f02c09cc6633158ed5b5d6cd7f1e
https://github.com/MohammedHAlali/pytorch_DoReFaNet/tree/d208089b9172f02c09cc6633158ed5b5d6cd7f1e
import torch import torch.utils.data import torch.nn as nn def uniform_quantize(k): class qfn(torch.autograd.Function): @staticmethod def forward(ctx, input): if k == 32: out = input elif k == 1: out = torch.sign(input) else: ...
RBFLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class RBFLayer(nn.Module): """ Transforms incoming data using a given radial basis function: u_{i} = rbf(||x - c_{i}|| / s_{i}) Arguments: in_features: size...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math from torch import nn import torch.nn.parallel import torch.opt...
MorganeAyle/SNIP-it
RBFLayer
false
854
[ "MIT" ]
0
df2bf44d6d3f7e4ea7733242a79c916735a7b49e
https://github.com/MorganeAyle/SNIP-it/tree/df2bf44d6d3f7e4ea7733242a79c916735a7b49e
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): """ Transforms incoming data using a given radial basis function: u_{i} = rbf(||x - c_{i}|| / s_{i}) Arguments: in_features: size of...
ChannelMixer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn.fun...
Misuzu-Kurenai/mlp-singer
ChannelMixer
false
855
[ "MIT" ]
0
416451045bb9b3965aaf496e84a8b45332a6ba59
https://github.com/Misuzu-Kurenai/mlp-singer/tree/416451045bb9b3965aaf496e84a8b45332a6ba59
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
TokenMixer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn.fun...
Misuzu-Kurenai/mlp-singer
TokenMixer
false
856
[ "MIT" ]
0
416451045bb9b3965aaf496e84a8b45332a6ba59
https://github.com/Misuzu-Kurenai/mlp-singer/tree/416451045bb9b3965aaf496e84a8b45332a6ba59
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
MultiHeadBoxAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductWithBoxAttention(nn.Module): """ Scaled dot-product attention with box """ def __init__(self, d_model, d_k, d_v, h, dropout=0.1, comment=None): """ :par...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Maxi-0902/DRAN
MultiHeadBoxAttention
false
857
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
from torch.nn import Module import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductWithBoxAttention(nn.Module): """ Scaled dot-product attention with box """ def __init__(self, d_model, d_k, d_v, h, dropout=0.1, comment=None): """ :par...
FixedSubnetConv
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch from torch import nn from torch.nn import functional as F import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class FixedSubnetConv(nn.Conv2d): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self....
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import math from torch import nn import torch.nn.parallel import torch.optim imp...
MorganeAyle/SNIP-it
FixedSubnetConv
false
858
[ "MIT" ]
0
df2bf44d6d3f7e4ea7733242a79c916735a7b49e
https://github.com/MorganeAyle/SNIP-it/tree/df2bf44d6d3f7e4ea7733242a79c916735a7b49e
import math import torch from torch import nn from torch.nn import functional as F import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Conv2d): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.scores = n...
SplitChannels
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch class SplitChannels(torch.nn.Module): def __init__(self, split_location): super(SplitChannels, self).__init__() self.split_location = split_location def forward(self, x): a, b = x[:, :self.split_location], x[:, self.split_location:] a, b = a.clone(), b.clone() ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda @triton.j...
Myyyr/segmentation
SplitChannels
false
859
[ "MIT" ]
0
6b9423e327cff1eb23599404031b7fb8e9ecf75d
https://github.com/Myyyr/segmentation/tree/6b9423e327cff1eb23599404031b7fb8e9ecf75d
import torch class Model(torch.nn.Module): def __init__(self, split_location): super().__init__() self.split_location = split_location def forward(self, x): a, b = x[:, :self.split_location], x[:, self.split_location:] a, b = a.clone(), b.clone() del x return ...
PoseMap
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class PoseMap(nn.Module): def __init__(self): super(PoseMap, self).__init__() pass def forward(self, x): assert len(x.shape) == 4, 'The HeatMap shape should be BxCxHxW' res = x.sum(dim=1, keepdim=True) H = x.shape[2] W = x.sh...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
MrChenFeng/Project-Template
PoseMap
false
860
[ "MIT" ]
0
42a335c6abb710bbae6407cbb0ca461533bc12f9
https://github.com/MrChenFeng/Project-Template/tree/42a335c6abb710bbae6407cbb0ca461533bc12f9
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() pass def forward(self, x): assert len(x.shape) == 4, 'The HeatMap shape should be BxCxHxW' res = x.sum(dim=1, keepdim=True) H = x.shape[2] W = x.shape[3] ...
ScalarAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.utils.data import torch.utils.checkpoint class ScalarAttention(nn.Module): def __init__(self, in_size, hidden_size): super(ScalarAttention, self).__init__() self.hidden = nn.Linear(in_size, hidden_size) nn.init.orthogonal_(self.hidden.weight...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
MarvinLvn/platalea
ScalarAttention
false
861
[ "Apache-2.0" ]
0
31def0813c90a3259f86f7d86cb576cd66dca3fe
https://github.com/MarvinLvn/platalea/tree/31def0813c90a3259f86f7d86cb576cd66dca3fe
import torch import torch.nn as nn import torch.utils.data import torch.utils.checkpoint class Model(nn.Module): def __init__(self, in_size, hidden_size): super().__init__() self.hidden = nn.Linear(in_size, hidden_size) nn.init.orthogonal_(self.hidden.weight.data) self.out = nn.Li...
MultiHeadAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductAttention(nn.Module): """ Scaled dot-product attention """ def __init__(self, d_model, d_k, d_v, h): """ :param d_model: Output dimensionality of the mo...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Maxi-0902/DRAN
MultiHeadAttention
false
862
[ "MIT" ]
0
c3dbfcbc018446544150dc4e151442d6a9fcd4d9
https://github.com/Maxi-0902/DRAN/tree/c3dbfcbc018446544150dc4e151442d6a9fcd4d9
from torch.nn import Module import torch import numpy as np import torch.nn as nn from torch.autograd import * class ScaledDotProductAttention(nn.Module): """ Scaled dot-product attention """ def __init__(self, d_model, d_k, d_v, h): """ :param d_model: Output dimensionality of the mo...
MNIST_classifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class MNIST_classifier(nn.Module): def __init__(self): super(MNIST_classifier, self).__init__() self.conv1 = nn.Conv2d(1, 32, 5, stride=2) self.conv2 =...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn import t...
MorganeAyle/SNIP-it
MNIST_classifier
false
863
[ "MIT" ]
0
df2bf44d6d3f7e4ea7733242a79c916735a7b49e
https://github.com/MorganeAyle/SNIP-it/tree/df2bf44d6d3f7e4ea7733242a79c916735a7b49e
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(1, 32, 5, stride=2) self.conv2 = nn.Conv2d(32, 64, 3, stride=2) ...
QuickGELU
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.distributed.nn class QuickGELU(nn.Module): def forward(self, x: 'torch.Tensor'): return x * torch.sigmoid(1.702 * x) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], {}]
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn import torch.distributed.nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch....
NYU-DICE-Lab/open_clip
QuickGELU
false
864
[ "MIT" ]
0
fd71804b503135fb1c7cc8de3a0d6599741c8ed9
https://github.com/NYU-DICE-Lab/open_clip/tree/fd71804b503135fb1c7cc8de3a0d6599741c8ed9
import torch from torch import nn import torch.distributed.nn class Model(nn.Module): def forward(self, x: 'torch.Tensor'): return x * torch.sigmoid(1.702 * x) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []
QREmbeddingBag
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn as nn from torch.nn.parameter import Parameter import torch.nn.functional as F class QREmbeddingBag(nn.Module): """Computes sums or means over two 'bags' of embeddings, one using the quotient of the indices and the other using the remainder of the indices, witho...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import numpy as np import torch.nn as nn from torch.nn.parameter import Paramet...
MrDoghead/dlrm
QREmbeddingBag
false
865
[ "MIT" ]
0
9b0d8ea992daa515104c7967f30110684283ebb1
https://github.com/MrDoghead/dlrm/tree/9b0d8ea992daa515104c7967f30110684283ebb1
import torch import numpy as np import torch.nn as nn from torch.nn.parameter import Parameter import torch.nn.functional as F class Model(nn.Module): """Computes sums or means over two 'bags' of embeddings, one using the quotient of the indices and the other using the remainder of the indices, without in...
MixerBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn.fun...
Misuzu-Kurenai/mlp-singer
MixerBlock
false
866
[ "MIT" ]
0
416451045bb9b3965aaf496e84a8b45332a6ba59
https://github.com/Misuzu-Kurenai/mlp-singer/tree/416451045bb9b3965aaf496e84a8b45332a6ba59
import torch import torch.nn.functional as F from torch import nn class FeedForward(nn.Module): def __init__(self, num_features, expansion_factor, dropout): super().__init__() num_hidden = expansion_factor * num_features self.fc1 = nn.Linear(num_features, num_hidden) self.fc2 = nn...
linformerAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn def get_EF(input_size, dim, method='learnable', head_dim=None, bias=True): """ Retuns the E or F matrix, initialized via xavier initialization. This is the recommended way to do it according to the authors of the paper. Includes a method for convolution, as well as a ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
MohammadrezaRezvani/performer-pytorch
linformerAttention
false
867
[ "MIT" ]
0
347dd58111f4f79b8991f7609552203609856b4b
https://github.com/MohammadrezaRezvani/performer-pytorch/tree/347dd58111f4f79b8991f7609552203609856b4b
import torch from torch import nn def get_EF(input_size, dim, method='learnable', head_dim=None, bias=True): """ Retuns the E or F matrix, initialized via xavier initialization. This is the recommended way to do it according to the authors of the paper. Includes a method for convolution, as well as a ...
FGFunction
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from torch.nn import functional as F class FGFunction(nn.Module): """Module used for F and G Archi : conv -> BN -> ReLu -> conv -> BN -> ReLu """ def __init__(self, channels): super(FGFunction, self).__init__() self.gn1 = nn.GroupNorm(1, channels, eps=0....
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import n...
Myyyr/segmentation
FGFunction
false
868
[ "MIT" ]
0
6b9423e327cff1eb23599404031b7fb8e9ecf75d
https://github.com/Myyyr/segmentation/tree/6b9423e327cff1eb23599404031b7fb8e9ecf75d
import torch from torch import nn from torch.nn import functional as F class Model(nn.Module): """Module used for F and G Archi : conv -> BN -> ReLu -> conv -> BN -> ReLu """ def __init__(self, channels): super().__init__() self.gn1 = nn.GroupNorm(1, channels, eps=0.001) self.con...
GLU
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.parallel import torch.onnx class GLU(nn.Module): def __init__(self): super(GLU, self).__init__() def forward(self, x): nc = x.size(1) assert nc % 2 == 0, 'channels dont divide 2!' nc = int(nc / 2) return x[:, :nc] * t...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.parallel import torch.onnx assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strid...
Nakachi-S/AttnGAN
GLU
false
869
[ "MIT" ]
0
2dfd1e38f78f2a58895d81131cd8c17e74dbacb2
https://github.com/Nakachi-S/AttnGAN/tree/2dfd1e38f78f2a58895d81131cd8c17e74dbacb2
import torch import torch.nn as nn import torch.nn.parallel import torch.onnx class Model(nn.Module): def __init__(self): super().__init__() def forward(self, x): nc = x.size(1) assert nc % 2 == 0, 'channels dont divide 2!' nc = int(nc / 2) return x[:, :nc] * torch.si...
LinearBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from scipy.stats import truncnorm def truncated_normal_(tensor, mean=0.0, std=1.0): values = truncnorm.rvs(-2, 2, size=tensor.shape) values = mean + std * values tensor.copy_(torch.from_numpy(values)) return tensor def fc_init_(module): if hasattr(module, 'weigh...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
JasonMa2016/learn2learn
LinearBlock
false
870
[ "MIT" ]
0
502e1ea6db64481d7464fdda4d4d0be9b0f1089a
https://github.com/JasonMa2016/learn2learn/tree/502e1ea6db64481d7464fdda4d4d0be9b0f1089a
import torch from torch import nn from scipy.stats import truncnorm def truncated_normal_(tensor, mean=0.0, std=1.0): values = truncnorm.rvs(-2, 2, size=tensor.shape) values = mean + std * values tensor.copy_(torch.from_numpy(values)) return tensor def fc_init_(module): if hasattr(module, 'weigh...
ClipLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn from torch.nn import functional as F from torch import distributed as dist import torch.distributed.nn def gather_features(image_features, text_features, local_loss=False, gather_with_grad=False, rank=0, world_size=1, use_horovod=False): if use_horovod: assert hvd is ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NYU-DICE-Lab/open_clip
ClipLoss
false
871
[ "MIT" ]
0
fd71804b503135fb1c7cc8de3a0d6599741c8ed9
https://github.com/NYU-DICE-Lab/open_clip/tree/fd71804b503135fb1c7cc8de3a0d6599741c8ed9
import torch from torch import nn from torch.nn import functional as F from torch import distributed as dist import torch.distributed.nn def gather_features(image_features, text_features, local_loss=False, gather_with_grad=False, rank=0, world_size=1, use_horovod=False): if use_horovod: assert hvd is ...
DenseBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class CausalConv1d(nn.Module): """A 1D causal convolution layer. Input: (B, D_in, T), where B is the minibatch size, D_in is the number of dimensions per step, and T is the number of steps. Output: (B, D_out, T), where B is the mi...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
NagisaZj/oyster
DenseBlock
false
872
[ "MIT" ]
0
069a510fe63bb29ecd9871e0e189e58b03c8cad9
https://github.com/NagisaZj/oyster/tree/069a510fe63bb29ecd9871e0e189e58b03c8cad9
import torch import torch.nn as nn import torch.nn.functional as F class CausalConv1d(nn.Module): """A 1D causal convolution layer. Input: (B, D_in, T), where B is the minibatch size, D_in is the number of dimensions per step, and T is the number of steps. Output: (B, D_out, T), where B is the mi...
distLinear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.utils.data from torch.nn.utils.weight_norm import WeightNorm import torch.nn.parallel import torch.optim class distLinear(nn.Module): def __init__(self, indim, outdim): super(distLinear, self).__init__() self.L = nn.Linear(indim, outdim, bias=False)...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
MuawizChaudhary/STARTUP
distLinear
false
873
[ "MIT" ]
0
03f39b34a4ec232f132173b4a1e67ea04165e52b
https://github.com/MuawizChaudhary/STARTUP/tree/03f39b34a4ec232f132173b4a1e67ea04165e52b
import torch import torch.nn as nn import torch.utils.data from torch.nn.utils.weight_norm import WeightNorm import torch.nn.parallel import torch.optim class Model(nn.Module): def __init__(self, indim, outdim): super().__init__() self.L = nn.Linear(indim, outdim, bias=False) self.class_w...
BothContextGate
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
NaomiatLibrary/OpenNMT-kpg-release
BothContextGate
false
874
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
SourceContextGate
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
NaomiatLibrary/OpenNMT-kpg-release
SourceContextGate
false
875
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
IIDTransform
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn.parallel import torch.utils.data from torchvision import transforms import torch.nn as nn import torch.cuda class IIDTransform(nn.Module): def __init__(self): super(IIDTransform, self).__init__() self.transform_op = transforms.Normalize((0.5,), (0.5,)) def mask_f...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn.parallel import torch.utils.data from torchvision import transforms impor...
NeilDG/NeuralNets-Experiment3
IIDTransform
false
876
[ "MIT" ]
0
f0d2f788eeca49f803f65810c155491ce687cf9e
https://github.com/NeilDG/NeuralNets-Experiment3/tree/f0d2f788eeca49f803f65810c155491ce687cf9e
import torch import torch.nn.parallel import torch.utils.data from torchvision import transforms import torch.nn as nn import torch.cuda class Model(nn.Module): def __init__(self): super().__init__() self.transform_op = transforms.Normalize((0.5,), (0.5,)) def mask_fill_nonzeros(self, input_...
TargetContextGate
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
NaomiatLibrary/OpenNMT-kpg-release
TargetContextGate
false
877
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
UpSampler
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from torch.nn import functional as F class UpSampler(nn.Module): """Up Sample module Decrease the channels size and increase the spatial size of tensor Extends: nn.Module """ def __init__(self, inChannels, outChannels, spatial_size): """...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
Myyyr/segmentation
UpSampler
false
878
[ "MIT" ]
0
6b9423e327cff1eb23599404031b7fb8e9ecf75d
https://github.com/Myyyr/segmentation/tree/6b9423e327cff1eb23599404031b7fb8e9ecf75d
import torch from torch import nn from torch.nn import functional as F class Model(nn.Module): """Up Sample module Decrease the channels size and increase the spatial size of tensor Extends: nn.Module """ def __init__(self, inChannels, outChannels, spatial_size): """ ...
Conv
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from math import sqrt assert_size_stride = torch._C._dynam...
NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio
Conv
false
879
[ "MIT" ]
0
231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
https://github.com/NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio/tree/231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
AverageAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.cuda import torch.distributed class PositionwiseFeedForward(nn.Module): """ A two-layer Feed-Forward-Network with residual layer norm. Args: d_model (int): the size of input for the first-layer of the FFN. d_ff (int): the hidden layer size of th...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.cuda import torch.distributed assert_size_str...
NaomiatLibrary/OpenNMT-kpg-release
AverageAttention
false
880
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class PositionwiseFeedForward(nn.Module): """ A two-layer Feed-Forward-Network with residual layer norm. Args: d_model (int): the size of input for the first-layer of the FFN. d_ff (int): the hidden layer size of th...
ContextGate
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.cuda import torch.distributed class ContextGate(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select t...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.cuda import torch.distributed assert_size_str...
NaomiatLibrary/OpenNMT-kpg-release
ContextGate
false
881
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class Model(nn.Module): """ Context gate is a decoder module that takes as input the previous word embedding, the current decoder state and the attention state, and produces a gate. The gate can be used to select the inp...
MultiHeadAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class MultiHeadAttention(nn.Module): def __init__(self, embedding_size, number_of_heads): super(MultiHeadAttention, self).__init__() self.embedding_size = embedding_size self.number_of_heads = number_of_heads self.head_dimension = embedding_size ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NMT-hub/transformer
MultiHeadAttention
false
882
[ "MIT" ]
0
e5b332da6a322e8025c30ee7e31fe34a323e7388
https://github.com/NMT-hub/transformer/tree/e5b332da6a322e8025c30ee7e31fe34a323e7388
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, embedding_size, number_of_heads): super().__init__() self.embedding_size = embedding_size self.number_of_heads = number_of_heads self.head_dimension = embedding_size // number_of_heads assert sel...
AdaIN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio
AdaIN
false
883
[ "MIT" ]
0
231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
https://github.com/NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio/tree/231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
GlobalAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.cuda import torch.distributed def aeq(*args): """ Assert all arguments have the same value """ arguments = (arg for arg in args) first = next(arguments) assert all(arg == first for arg in arguments ), 'Not ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NaomiatLibrary/OpenNMT-kpg-release
GlobalAttention
false
884
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.nn.functional as F import torch.cuda import torch.distributed def aeq(*args): """ Assert all arguments have the same value """ arguments = (arg for arg in args) first = next(arguments) assert all(arg == first for arg in arguments ), 'Not ...
PolicyNetwork
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class PolicyNetwork(torch.nn.Module): def __init__(self, input_size, output_size): super().__init__() self.input_size = input_size self.linear1 = nn.Linear(input_size, 32) self.linear2 = nn.Linear(32, output_size) ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NiccoloSacchi/rlcard
PolicyNetwork
false
885
[ "MIT" ]
0
046129e8616b12e25652957869a94ab5fd838ae1
https://github.com/NiccoloSacchi/rlcard/tree/046129e8616b12e25652957869a94ab5fd838ae1
import torch import torch.nn as nn import torch.nn.functional as F class Model(torch.nn.Module): def __init__(self, input_size, output_size): super().__init__() self.input_size = input_size self.linear1 = nn.Linear(input_size, 32) self.linear2 = nn.Linear(32, output_size) def...
VdConv1D
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn as nn import torch.nn.functional as F def calculate_kl(log_alpha): return 0.5 * torch.sum(torch.log1p(torch.exp(-log_alpha))) class VdConv1D(nn.Module): """ Conv1D Layer variational dropout """ def __init__(self, in_channels, out_channels, kernel_size, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math im...
Neronjust2017/pytorch-classification-project
VdConv1D
false
886
[ "MIT" ]
0
fc5f4d7c46d071765f682ce20e6580646d4e5c76
https://github.com/Neronjust2017/pytorch-classification-project/tree/fc5f4d7c46d071765f682ce20e6580646d4e5c76
import math import torch import torch.nn as nn import torch.nn.functional as F def calculate_kl(log_alpha): return 0.5 * torch.sum(torch.log1p(torch.exp(-log_alpha))) class Model(nn.Module): """ Conv1D Layer variational dropout """ def __init__(self, in_channels, out_channels, kernel_size, alp...
SpatialAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.parallel import torch.utils.data import torch.nn as nn import torch.cuda class SpatialAttention(nn.Module): def __init__(self): super(SpatialAttention, self).__init__() self.conv = nn.Conv2d(2, 1, 7, padding=3, bias=False) self.sigmoid = nn.Sigmoid() def ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn.parallel impo...
NeilDG/NeuralNets-Experiment3
SpatialAttention
false
887
[ "MIT" ]
0
f0d2f788eeca49f803f65810c155491ce687cf9e
https://github.com/NeilDG/NeuralNets-Experiment3/tree/f0d2f788eeca49f803f65810c155491ce687cf9e
import torch import torch.nn.parallel import torch.utils.data import torch.nn as nn import torch.cuda class Model(nn.Module): def __init__(self): super().__init__() self.conv = nn.Conv2d(2, 1, 7, padding=3, bias=False) self.sigmoid = nn.Sigmoid() def forward(self, x: 'torch.Tensor') ...
Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import math import torch import torch.nn as nn class Loss(nn.Module): def __init__(self, device, type_in='pred_intervals', alpha=0.1, loss_type='qd_soft', censor_R=False, soften=100.0, lambda_in=10.0, sigma_in=0.5): super().__init__() self.alpha = alpha self.lambda_in = la...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
Neronjust2017/pytorch-classification-project
Loss
false
888
[ "MIT" ]
0
fc5f4d7c46d071765f682ce20e6580646d4e5c76
https://github.com/Neronjust2017/pytorch-classification-project/tree/fc5f4d7c46d071765f682ce20e6580646d4e5c76
import math import torch import torch.nn as nn class Model(nn.Module): def __init__(self, device, type_in='pred_intervals', alpha=0.1, loss_type='qd_soft', censor_R=False, soften=100.0, lambda_in=10.0, sigma_in=0.5): super().__init__() self.alpha = alpha self.lambda_in = l...
Linear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from math import sqrt assert_size_stride = torch._C._dynam...
NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio
Linear
false
889
[ "MIT" ]
0
231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
https://github.com/NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio/tree/231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
PixelNorm
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class PixelNorm(nn.Module): def __init__(self): super().__init__() def forward(self, input): return input / torch.sqrt(torch.mean(input ** 2, dim=0, keepdim= True) + 1e-08) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio
PixelNorm
false
890
[ "MIT" ]
0
231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
https://github.com/NethraGunti/Woven-Artificial-Profile-WARP-Face-Video-Synthesis-from-Profile-and-Audio/tree/231d8daa8dddfd5eda8a092eb99c5d0e59d8b3f7
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, input): return input / torch.sqrt(torch.mean(input ** 2, dim=0, keepdim= True) + 1e-08) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inp...
LayerNorm
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class LayerNorm(nn.Module): """Construct a layernorm module in the OpenAI style (epsilon inside the square root).""" def __init__(self, n_state, e=1e-05): super(LayerNorm, self).__init__() self.g = nn.Parameter(torch.ones(n_state)) self.b = nn.Parame...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
NickSchoelkopf/SummerTime
LayerNorm
false
891
[ "Apache-2.0" ]
0
9a89aab8e1544e3c52c043b9c47ab325e665e11e
https://github.com/NickSchoelkopf/SummerTime/tree/9a89aab8e1544e3c52c043b9c47ab325e665e11e
import torch import torch.nn as nn class Model(nn.Module): """Construct a layernorm module in the OpenAI style (epsilon inside the square root).""" def __init__(self, n_state, e=1e-05): super().__init__() self.g = nn.Parameter(torch.ones(n_state)) self.b = nn.Parameter(torch.zeros(n_s...
VdLinear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn as nn import torch.nn.functional as F def calculate_kl(log_alpha): return 0.5 * torch.sum(torch.log1p(torch.exp(-log_alpha))) class VdLinear(nn.Module): """ Linear Layer variational dropout """ def __init__(self, n_in, n_out, alpha_shape=(1, 1), bias=Tr...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math im...
Neronjust2017/pytorch-classification-project
VdLinear
false
892
[ "MIT" ]
0
fc5f4d7c46d071765f682ce20e6580646d4e5c76
https://github.com/Neronjust2017/pytorch-classification-project/tree/fc5f4d7c46d071765f682ce20e6580646d4e5c76
import math import torch import torch.nn as nn import torch.nn.functional as F def calculate_kl(log_alpha): return 0.5 * torch.sum(torch.log1p(torch.exp(-log_alpha))) class Model(nn.Module): """ Linear Layer variational dropout """ def __init__(self, n_in, n_out, alpha_shape=(1, 1), bias=True)...
C
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class C(nn.Module): def __init__(self, input_channel, output_channel, kernel_size, stride, padding, activation=None): """ At the final layer, a 3x3 convolution is used to map each 64-component feature vector to the desired number of classes. ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
Nikronic/Deep-Halftoning
C
false
893
[ "MIT" ]
0
9564c592abf139ccab2791c1dbb354505edab5f9
https://github.com/Nikronic/Deep-Halftoning/tree/9564c592abf139ccab2791c1dbb354505edab5f9
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, input_channel, output_channel, kernel_size, stride, padding, activation=None): """ At the final layer, a 3x3 convolution is used to map each 64-component feature vector to the desired number of classes. ...
ConcatReLU
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F def concat_relu(x): """Concatenated ReLU (http://arxiv.org/abs/1603.05201).""" return F.relu(torch.cat([x, -x], dim=1)) class ConcatReLU(nn.Module): """Concatenated ReLU (http://arxiv.org/abs/1603.05201).""" def forward(self, input)...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import torch.nn.functional as F assert_size_stride = torch._C._dyna...
Nintorac/survae_experiments
ConcatReLU
false
894
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import torch import torch.nn as nn import torch.nn.functional as F def concat_relu(x): """Concatenated ReLU (http://arxiv.org/abs/1603.05201).""" return F.relu(torch.cat([x, -x], dim=1)) class Model(nn.Module): """Concatenated ReLU (http://arxiv.org/abs/1603.05201).""" def forward(self, input): ...
CrossEntropyLossOneHot
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import Tensor from torch.nn.modules.loss import CrossEntropyLoss class CrossEntropyLossOneHot(CrossEntropyLoss): EPS: 'int' = 1e-07 def forward(self, input: 'Tensor', target: 'Tensor') ->Tensor: assert self.weight is None or isinstance(self.weight, Tensor) input = torc...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from torch.nn.modules....
NikolayZakharevich/music-processing
CrossEntropyLossOneHot
false
895
[ "MIT" ]
0
516a3bca585f211d232cac7ede6cc417fb8878fe
https://github.com/NikolayZakharevich/music-processing/tree/516a3bca585f211d232cac7ede6cc417fb8878fe
import torch from torch import Tensor from torch.nn.modules.loss import CrossEntropyLoss class Model(CrossEntropyLoss): EPS: 'int' = 1e-07 def forward(self, input: 'Tensor', target: 'Tensor') ->Tensor: assert self.weight is None or isinstance(self.weight, Tensor) input = torch.clip(input, sel...
Lookahead
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.utils.data.distributed from torch import nn import torch.nn.functional as F class Lookahead(nn.Module): def __init__(self, n_features, context): super(Lookahead, self).__init__() assert context > 0 self.context = context self.n_features = n_features ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.utils.data.distributed from torch import nn assert_size_stride = to...
NikolaiBabkin/deepspeech.pytorch
Lookahead
false
896
[ "MIT" ]
0
2b120c6b735cc46200e10f81e169c8d7b75e8495
https://github.com/NikolaiBabkin/deepspeech.pytorch/tree/2b120c6b735cc46200e10f81e169c8d7b75e8495
import torch import torch.utils.data.distributed from torch import nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, n_features, context): super().__init__() assert context > 0 self.context = context self.n_features = n_features self.pad = 0, se...
ResidualAttentionBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from typing import Callable from torch import nn from torch.nn import functional as F import torch.distributed.nn from collections import OrderedDict from typing import Optional class LayerNorm(nn.LayerNorm): """Subclass torch's LayerNorm to handle fp16.""" def forward(self, x: 'torch.Tensor'): ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NYU-DICE-Lab/open_clip
ResidualAttentionBlock
false
897
[ "MIT" ]
0
fd71804b503135fb1c7cc8de3a0d6599741c8ed9
https://github.com/NYU-DICE-Lab/open_clip/tree/fd71804b503135fb1c7cc8de3a0d6599741c8ed9
import torch from typing import Callable from torch import nn from torch.nn import functional as F import torch.distributed.nn from collections import OrderedDict from typing import Optional class LayerNorm(nn.LayerNorm): """Subclass torch's LayerNorm to handle fp16.""" def forward(self, x: 'torch.Tensor'): ...
SoftCrossEntropyLoss2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data.distributed import torch import torch.nn as nn from numpy import int64 as int64 from torchvision.transforms import functional as F import torch.nn.functional as F import torch.utils class SoftCrossEntropyLoss2d(nn.Module): def __init__(self): super(SoftCrossEntropyLos...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
HRTNet/HRTNet
SoftCrossEntropyLoss2d
false
898
[ "MIT" ]
0
6a51c9c34568988ea6125a1638794c63d8fadbea
https://github.com/HRTNet/HRTNet/tree/6a51c9c34568988ea6125a1638794c63d8fadbea
import torch import torch.utils.data.distributed import torch import torch.nn as nn from numpy import int64 as int64 from torchvision.transforms import functional as F import torch.nn.functional as F import torch.utils class Model(nn.Module): def __init__(self): super().__init__() def forward(self, ...
GatedTanhUnit
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn def gated_tanh(x, dim): """Gated Tanh activation.""" x_tanh, x_sigmoid = torch.chunk(x, 2, dim=dim) return torch.tanh(x_tanh) * torch.sigmoid(x_sigmoid) class GatedTanhUnit(nn.Module): """Gated Tanh activation.""" def __init__(self, dim=-1): super(Gate...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
Nintorac/survae_experiments
GatedTanhUnit
false
899
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import torch import torch.nn as nn def gated_tanh(x, dim): """Gated Tanh activation.""" x_tanh, x_sigmoid = torch.chunk(x, 2, dim=dim) return torch.tanh(x_tanh) * torch.sigmoid(x_sigmoid) class Model(nn.Module): """Gated Tanh activation.""" def __init__(self, dim=-1): super().__init__()...
ConcatELU
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F def concat_elu(x): """Like concatenated ReLU (http://arxiv.org/abs/1603.05201), but with ELU instead.""" return F.elu(torch.cat([x, -x], dim=1)) class ConcatELU(nn.Module): """Like concatenated ReLU (http://arxiv.org/abs/1603.05201), but...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn import torch.nn.functional as F assert_size_stride = torc...
Nintorac/survae_experiments
ConcatELU
false
900
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import torch import torch.nn as nn import torch.nn.functional as F def concat_elu(x): """Like concatenated ReLU (http://arxiv.org/abs/1603.05201), but with ELU instead.""" return F.elu(torch.cat([x, -x], dim=1)) class Model(nn.Module): """Like concatenated ReLU (http://arxiv.org/abs/1603.05201), but wit...
PositionalEncoding1d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn as nn class PositionalEncoding1d(nn.Module): """ Learning positional embeddings. Args: shape: Iterable, the shape of the input. embedding_dim: int, the size of each embedding vector. """ def __init__(self, size, embedding_dim): sup...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import math import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guar...
Nintorac/survae_experiments
PositionalEncoding1d
false
901
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import math import torch import torch.nn as nn class Model(nn.Module): """ Learning positional embeddings. Args: shape: Iterable, the shape of the input. embedding_dim: int, the size of each embedding vector. """ def __init__(self, size, embedding_dim): super().__init__()...
AutoregressiveShift
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class AutoregressiveShift(nn.Module): """Shifts input right to make model autoregressive.""" def __init__(self, embed_dim): super(AutoregressiveShift, self).__init__() self.embed_dim = embed_dim self.first_token = nn.Parameter(torch.Tensor(1, 1, embe...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Nintorac/survae_experiments
AutoregressiveShift
false
902
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import torch import torch.nn as nn class Model(nn.Module): """Shifts input right to make model autoregressive.""" def __init__(self, embed_dim): super().__init__() self.embed_dim = embed_dim self.first_token = nn.Parameter(torch.Tensor(1, 1, embed_dim)) self._reset_parameters(...
GatedConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class GatedConv2d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, padding): super(GatedConv2d, self).__init__() self.in_channels = in_channels self.conv = nn.Conv2d(in_channels, out_channels * 3, kernel_size= kernel_siz...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
Nintorac/survae_experiments
GatedConv2d
false
903
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, padding): super().__init__() self.in_channels = in_channels self.conv = nn.Conv2d(in_channels, out_channels * 3, kernel_size= kernel_size, padding=padding) ...
HeatmapLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data class HeatmapLoss(torch.nn.Module): """ loss for detection heatmap """ def __init__(self): super(HeatmapLoss, self).__init__() def forward(self, pred, gt): l = (pred - gt) ** 2 l = l.mean(dim=3).mean(dim=2).mean(dim=1) return l...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_...
NiranthS/pytorch_stacked_hourglass
HeatmapLoss
false
904
[ "BSD-3-Clause" ]
0
db9838eb13f6848ba3b9db844c1e023eb8688c3c
https://github.com/NiranthS/pytorch_stacked_hourglass/tree/db9838eb13f6848ba3b9db844c1e023eb8688c3c
import torch import torch.utils.data class Model(torch.nn.Module): """ loss for detection heatmap """ def __init__(self): super().__init__() def forward(self, pred, gt): l = (pred - gt) ** 2 l = l.mean(dim=3).mean(dim=2).mean(dim=1) return l def get_inputs(): ...
AvgPoolHead
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.optim class AvgPoolHead(nn.Module): def __init__(self, in_channels, out_channels, fea_map_size): super(AvgPoolHead, self).__init__() self.avgpool = nn.AvgPool2d(fea_map_size, stride=1) self.fc = nn.Linear(in_channels, out_channels) def ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.optim assert_size_stride = torch._C._dynamo.g...
NiteshBharadwaj/structured_aleatoric_uncertainty_for_human_pose
AvgPoolHead
false
905
[ "MIT" ]
0
c74fb7384be562f0a0f1966b3fadf19e13a235f2
https://github.com/NiteshBharadwaj/structured_aleatoric_uncertainty_for_human_pose/tree/c74fb7384be562f0a0f1966b3fadf19e13a235f2
import torch import torch.nn as nn import torch.optim class Model(nn.Module): def __init__(self, in_channels, out_channels, fea_map_size): super().__init__() self.avgpool = nn.AvgPool2d(fea_map_size, stride=1) self.fc = nn.Linear(in_channels, out_channels) def forward(self, x): ...
PositionalEncodingImage
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn as nn class PositionalEncodingImage(nn.Module): """ Learning positional embeddings for images. Embeddings for channel, height and width are added to form the full positional embedding. These encodings correspond to the ones from Sparse Transformers (https://arx...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import math import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guar...
Nintorac/survae_experiments
PositionalEncodingImage
false
906
[ "MIT" ]
0
d68cc25e2604aab08b53617c1f3ffe4716f166c4
https://github.com/Nintorac/survae_experiments/tree/d68cc25e2604aab08b53617c1f3ffe4716f166c4
import math import torch import torch.nn as nn class Model(nn.Module): """ Learning positional embeddings for images. Embeddings for channel, height and width are added to form the full positional embedding. These encodings correspond to the ones from Sparse Transformers (https://arxiv.org/abs/1904.10...
LinearZeros
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class LinearZeros(nn.Linear): def __init__(self, in_features, out_features, bias=True, logscale_factor=3.0): """ Linear layer with zero initialization :param in_features: size of each input sample :type in_features: int :param ou...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch....
NirDiamant/pytorch-glow
LinearZeros
false
907
[ "MIT" ]
0
2ab11f3a8486b86a279fe4fa64f25aa91226ee8a
https://github.com/NirDiamant/pytorch-glow/tree/2ab11f3a8486b86a279fe4fa64f25aa91226ee8a
import torch import torch.nn as nn class Model(nn.Linear): def __init__(self, in_features, out_features, bias=True, logscale_factor=3.0): """ Linear layer with zero initialization :param in_features: size of each input sample :type in_features: int :param out_feat...
Conv2dZeros
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ActNorm(nn.Module): def __init__(self, num_channels, scale=1.0, logscale_factor=3.0, batch_variance=False): """ Activation normalization layer :param num_channels: number of channels :type num_channels: int :param scale: sc...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch....
NirDiamant/pytorch-glow
Conv2dZeros
false
908
[ "MIT" ]
0
2ab11f3a8486b86a279fe4fa64f25aa91226ee8a
https://github.com/NirDiamant/pytorch-glow/tree/2ab11f3a8486b86a279fe4fa64f25aa91226ee8a
import torch import torch.nn as nn class ActNorm(nn.Module): def __init__(self, num_channels, scale=1.0, logscale_factor=3.0, batch_variance=False): """ Activation normalization layer :param num_channels: number of channels :type num_channels: int :param scale: sc...
GaussianKernel
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import Optional import torch.nn as nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class GaussianKernel(nn.Module): """Gaussian Kernel Matrix Gaussian Kernel k is defined by .. math:: k(x_1, x_2) = \\exp \\left( ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from typing import Opt...
NiteshBharadwaj/ignoringhumanpose
GaussianKernel
false
909
[ "MIT" ]
0
1fb7a063fded9cff18f7de4e1d71845983077256
https://github.com/NiteshBharadwaj/ignoringhumanpose/tree/1fb7a063fded9cff18f7de4e1d71845983077256
import torch from typing import Optional import torch.nn as nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): """Gaussian Kernel Matrix Gaussian Kernel k is defined by .. math:: k(x_1, x_2) = \\exp \\left( - \\dfrac...
Multihead_Attention_Layer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn as nn import torch.nn.functional as F def scaled_self_attention(q, k, v, key_size): weight = torch.matmul(q, k) weight = F.softmax(weight / math.sqrt(key_size), dim=-1) attention = torch.matmul(weight, v) return attention class Multihead_Attention_Layer(nn.Mo...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
NilsLusch/Point-Cloud-Transformer
Multihead_Attention_Layer
false
910
[ "MIT" ]
0
84a16b45b8949bbf8e7730b10bd5835e2ab4e642
https://github.com/NilsLusch/Point-Cloud-Transformer/tree/84a16b45b8949bbf8e7730b10bd5835e2ab4e642
import math import torch import torch.nn as nn import torch.nn.functional as F def scaled_self_attention(q, k, v, key_size): weight = torch.matmul(q, k) weight = F.softmax(weight / math.sqrt(key_size), dim=-1) attention = torch.matmul(weight, v) return attention class Model(nn.Module): def __in...
Conv
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Conv(nn.Module): """ Convolution Module """ def __init__(self, in_channels, out_channels, kernel_size=1, stride=1, padding=0, dilation=1, bias=True, w_init='linear'): """ :param in_channels: dimension of input :param out_channel...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
Numb523/FastSpeech2_emotion
Conv
false
911
[ "MIT" ]
0
a541ce89ddf66625ee57c0a294d0bec1ae701f0c
https://github.com/Numb523/FastSpeech2_emotion/tree/a541ce89ddf66625ee57c0a294d0bec1ae701f0c
import torch import torch.nn as nn class Model(nn.Module): """ Convolution Module """ def __init__(self, in_channels, out_channels, kernel_size=1, stride=1, padding=0, dilation=1, bias=True, w_init='linear'): """ :param in_channels: dimension of input :param out_channe...
Theta
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.autograd import Function import torch from typing import Optional from typing import Tuple import torch.nn as nn from typing import Any import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class GradientReverseFunction(Function): @staticmethod def...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch.autograd import Function from typing import Optional from typing impo...
NiteshBharadwaj/ignoringhumanpose
Theta
false
912
[ "MIT" ]
0
1fb7a063fded9cff18f7de4e1d71845983077256
https://github.com/NiteshBharadwaj/ignoringhumanpose/tree/1fb7a063fded9cff18f7de4e1d71845983077256
from torch.autograd import Function import torch from typing import Optional from typing import Tuple import torch.nn as nn from typing import Any import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class GradientReverseFunction(Function): @staticmethod def...
LastTimeStep
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.utils.data from typing import Tuple class LastTimeStep(nn.Module): """ A class for extracting the hidden activations of the last time step following the output of a PyTorch RNN module. """ def __init__(self, bidirectional=False): super(Last...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._...
Onion-Team-VN/skilledlab
LastTimeStep
false
913
[ "Apache-2.0" ]
0
ac5cd7b5aee52da98aee8a32e5d161fd8b7dddab
https://github.com/Onion-Team-VN/skilledlab/tree/ac5cd7b5aee52da98aee8a32e5d161fd8b7dddab
import torch from torch import nn import torch.utils.data from typing import Tuple class Model(nn.Module): """ A class for extracting the hidden activations of the last time step following the output of a PyTorch RNN module. """ def __init__(self, bidirectional=False): super().__init__(...
myLoss2
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn.functional as F import torch.nn as nn class myLoss2(nn.Module): def __init__(self, alpha=1.0): super(myLoss2, self).__init__() self.alpha = alpha def forward(self, sent_probs, doc_probs, sent_targets, doc_targets): loss_1 = F.mse_loss(sent_probs, sent_tar...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
PKULiuHui/LiveBlogSum
myLoss2
false
914
[ "MIT" ]
0
b6a22521ee454e649981d70ddca6c89a1bac5a4c
https://github.com/PKULiuHui/LiveBlogSum/tree/b6a22521ee454e649981d70ddca6c89a1bac5a4c
import torch import torch.nn.functional as F import torch.nn as nn class Model(nn.Module): def __init__(self, alpha=1.0): super().__init__() self.alpha = alpha def forward(self, sent_probs, doc_probs, sent_targets, doc_targets): loss_1 = F.mse_loss(sent_probs, sent_targets) l...
relu_constant_fraction
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import numpy as np from torch import nn from torch.nn.functional import relu def regula_falsi(func, a, b, iterations): f_a = func(a, -1) f_b = func(b, -1) if torch.any(f_a * f_b >= 0): None raise Exception( 'You have not assumed right initial values in regula falsi...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import numpy as np from torch import nn from torch.nn.functional import relu assert_size_...
Noppornying00/constant-fraction-activation
relu_constant_fraction
false
915
[ "Apache-2.0" ]
0
b25745e7339df13e3db34d8c8372d5cbaa3c13bb
https://github.com/Noppornying00/constant-fraction-activation/tree/b25745e7339df13e3db34d8c8372d5cbaa3c13bb
import torch import numpy as np from torch import nn from torch.nn.functional import relu def regula_falsi(func, a, b, iterations): f_a = func(a, -1) f_b = func(b, -1) if torch.any(f_a * f_b >= 0): None raise Exception( 'You have not assumed right initial values in regula falsi...
MarginDisparityDiscrepancy
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import Optional import torch.nn as nn import torch.nn.functional as F import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed def shift_log(x: 'torch.Tensor', offset: 'Optional[float]'=1e-06 ) ->torch.Tensor: """ First shift, then ca...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from typing import Opt...
NiteshBharadwaj/ignoringhumanpose
MarginDisparityDiscrepancy
false
916
[ "MIT" ]
0
1fb7a063fded9cff18f7de4e1d71845983077256
https://github.com/NiteshBharadwaj/ignoringhumanpose/tree/1fb7a063fded9cff18f7de4e1d71845983077256
import torch from typing import Optional import torch.nn as nn import torch.nn.functional as F import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed def shift_log(x: 'torch.Tensor', offset: 'Optional[float]'=1e-06 ) ->torch.Tensor: """ First shift, then ca...
DivisiveNormalization2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
from torch.nn import Module import torch from torch import Tensor from typing import Union from typing import Tuple import torch.nn.functional as F class DivisiveNormalization2d(Module): """Applies a 2D divisive normalization over an input signal composed of several input planes. In the simplest case, th...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from torch.nn import M...
Noppornying00/constant-fraction-activation
DivisiveNormalization2d
false
917
[ "Apache-2.0" ]
0
b25745e7339df13e3db34d8c8372d5cbaa3c13bb
https://github.com/Noppornying00/constant-fraction-activation/tree/b25745e7339df13e3db34d8c8372d5cbaa3c13bb
from torch.nn import Module import torch from torch import Tensor from typing import Union from typing import Tuple import torch.nn.functional as F class Model(Module): """Applies a 2D divisive normalization over an input signal composed of several input planes. In the simplest case, the output value of ...
UPChannelRPN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F def xcorr_fast(x, kernel): """group conv2d to calculate cross correlation, fast version """ batch = kernel.size()[0] pk = kernel.view(-1, x.size()[1], kernel.size()[2], kernel.size()[3]) px = x.view(1, -1, x.size()[2], x.size()[3])...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.functional as F assert_size_stride = torch...
LSH9832/MyPythonModules
UPChannelRPN
false
918
[ "MIT" ]
0
442566a0fbd6ebe2bc20b6914686a1e2663d10c0
https://github.com/LSH9832/MyPythonModules/tree/442566a0fbd6ebe2bc20b6914686a1e2663d10c0
import torch import torch.nn as nn import torch.nn.functional as F def xcorr_fast(x, kernel): """group conv2d to calculate cross correlation, fast version """ batch = kernel.size()[0] pk = kernel.view(-1, x.size()[1], kernel.size()[2], kernel.size()[3]) px = x.view(1, -1, x.size()[2], x.size()[3])...
MatrixTree
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.cuda import torch.distributed class MatrixTree(nn.Module): """Implementation of the matrix-tree theorem for computing marginals of non-projective dependency parsing. This attention layer is used in the paper "Learning Structured Text Representations" :ci...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn import torch.cuda import torch.distributed assert_s...
NaomiatLibrary/OpenNMT-kpg-release
MatrixTree
false
919
[ "MIT" ]
0
1da3468d7dad22529a77f3526abf9b373bd3dc4c
https://github.com/NaomiatLibrary/OpenNMT-kpg-release/tree/1da3468d7dad22529a77f3526abf9b373bd3dc4c
import torch import torch.nn as nn import torch.cuda import torch.distributed class Model(nn.Module): """Implementation of the matrix-tree theorem for computing marginals of non-projective dependency parsing. This attention layer is used in the paper "Learning Structured Text Representations" :cite:`D...
JointsMSELossNoReduction
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class JointsMSELossNoReduction(nn.Module): def __init__(self, use_target_weight, logger): super(JointsMSELossNoReduction, self).__init__() self.criterion = la...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed assert_size_st...
NiteshBharadwaj/ignoringhumanpose
JointsMSELossNoReduction
false
920
[ "MIT" ]
0
1fb7a063fded9cff18f7de4e1d71845983077256
https://github.com/NiteshBharadwaj/ignoringhumanpose/tree/1fb7a063fded9cff18f7de4e1d71845983077256
import torch import torch.nn as nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): def __init__(self, use_target_weight, logger): super().__init__() self.criterion = lambda x, y: ((x - y) ** 2).sum(1).unsqueeze(1) ...
Attention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import torch.nn import torch.optim from torch.nn import functional as F from torch import nn class Attention(nn.Module): def __init__(self, hidden_size): super(Attention, self).__init__() self.hidden_size = hidden_size self.attn = nn.Linear(self.hidden_size * 2, h...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
OneAdder/hseling-repo-chukchi-type
Attention
false
921
[ "MIT" ]
0
5f5e651510bca7cfb89dc2e98b07bcc63b6330a4
https://github.com/OneAdder/hseling-repo-chukchi-type/tree/5f5e651510bca7cfb89dc2e98b07bcc63b6330a4
import math import torch import torch.nn import torch.optim from torch.nn import functional as F from torch import nn class Model(nn.Module): def __init__(self, hidden_size): super().__init__() self.hidden_size = hidden_size self.attn = nn.Linear(self.hidden_size * 2, hidden_size) ...
myLoss3
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn.functional as F import torch.nn as nn class myLoss3(nn.Module): def __init__(self, alpha=1.0, beta=1.0): super(myLoss3, self).__init__() self.alpha = alpha self.beta = beta def forward(self, sent_probs, doc_probs, event_probs, sent_targets, doc_ta...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
PKULiuHui/LiveBlogSum
myLoss3
false
924
[ "MIT" ]
0
b6a22521ee454e649981d70ddca6c89a1bac5a4c
https://github.com/PKULiuHui/LiveBlogSum/tree/b6a22521ee454e649981d70ddca6c89a1bac5a4c
import torch import torch.nn.functional as F import torch.nn as nn class Model(nn.Module): def __init__(self, alpha=1.0, beta=1.0): super().__init__() self.alpha = alpha self.beta = beta def forward(self, sent_probs, doc_probs, event_probs, sent_targets, doc_targets, event_ta...
GraphLearner
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch from torch.nn.modules.module import Module import torch.nn as nn import torch.nn.functional as F class GraphLearner(Module): def __init__(self, in_feature_dim, combined_feature_dim, n_obj, dropout=0.0 ): """ eq(1): A=EE^T, build adj matrix ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Originofamonia/vqa-project
GraphLearner
false
925
[ "Apache-2.0" ]
0
cf67b62ddf5732881dfb4278478accfd15c4df6d
https://github.com/Originofamonia/vqa-project/tree/cf67b62ddf5732881dfb4278478accfd15c4df6d
from torch.nn import Module import torch from torch.nn.modules.module import Module import torch.nn as nn import torch.nn.functional as F class Model(Module): def __init__(self, in_feature_dim, combined_feature_dim, n_obj, dropout=0.0 ): """ eq(1): A=EE^T, build adj matrix ## Vari...
AE_Net
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch.optim import * from torch import nn class AE_Net(nn.Module): """docstring for AE_Net.""" def __init__(self, input_shape): super(AE_Net, self).__init__() self.encoder_hidden_layer = nn.Linear(in_features=input_shape, out_features=128) self.encoder_ou...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch.optim import * fro...
Paratra/IoTAnalytics_pub
AE_Net
false
926
[ "MIT" ]
0
8c1d02b60ef609c3cba654ce4a5568c39fc63edf
https://github.com/Paratra/IoTAnalytics_pub/tree/8c1d02b60ef609c3cba654ce4a5568c39fc63edf
import torch from torch.optim import * from torch import nn class Model(nn.Module): """docstring for AE_Net.""" def __init__(self, input_shape): super().__init__() self.encoder_hidden_layer = nn.Linear(in_features=input_shape, out_features=128) self.encoder_output_layer = ...
NormedLinear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F from torch import nn class NormedLinear(nn.Linear): """Normalized Linear Layer. Args: tempeature (float, optional): Tempeature term. Default to 20. power (int, optional): Power term. Default to 1.0. eps (float, optional): The minimal value ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import n...
Parskatt/mmdetection
NormedLinear
false
927
[ "Apache-2.0" ]
0
ee4cfa29e7f479b2454b1f1355f8c05be62d8466
https://github.com/Parskatt/mmdetection/tree/ee4cfa29e7f479b2454b1f1355f8c05be62d8466
import torch import torch.nn.functional as F from torch import nn class Model(nn.Linear): """Normalized Linear Layer. Args: tempeature (float, optional): Tempeature term. Default to 20. power (int, optional): Power term. Default to 1.0. eps (float, optional): The minimal value of divi...
NormedConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class NormedConv2d(nn.Conv2d): """Normalized Conv2d Layer. Args: tempeature (float, optional): Tempeature term. Default to 20. power (int, optional): Power term. Default to 1.0. eps (float, optional): The minimal value of divisor to keep ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import n...
Parskatt/mmdetection
NormedConv2d
false
928
[ "Apache-2.0" ]
0
ee4cfa29e7f479b2454b1f1355f8c05be62d8466
https://github.com/Parskatt/mmdetection/tree/ee4cfa29e7f479b2454b1f1355f8c05be62d8466
import torch from torch import nn class Model(nn.Conv2d): """Normalized Conv2d Layer. Args: tempeature (float, optional): Tempeature term. Default to 20. power (int, optional): Power term. Default to 1.0. eps (float, optional): The minimal value of divisor to keep numeric...
Net
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class Net(nn.Module): def __init__(self, num_rej=0): super(Net, self).__init__() self.num_rej = num_rej + 1 self.conv1 = nn.Conv2d(3, 6, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import ...
NlGG/Rejection
Net
false
929
[ "MIT" ]
0
5f7cc64b71dacc2eb794b3f7c48390457e363cc5
https://github.com/NlGG/Rejection/tree/5f7cc64b71dacc2eb794b3f7c48390457e363cc5
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class Model(nn.Module): def __init__(self, num_rej=0): super().__init__() self.num_rej = num_rej + 1 self.conv1 = nn.Conv2d(3, 6, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 = nn.Co...
BPMLLLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import Tensor class BPMLLLoss(torch.nn.Module): def __init__(self, bias=(1, 1)): super(BPMLLLoss, self).__init__() self.bias = bias assert len(self.bias) == 2 and all(map(lambda x: isinstance(x, int) and x > 0, bias)), 'bias must be positive integers' ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import Tensor assert_size_stride = torch._C._dynamo.guards.assert_size_stride ...
PaulStryck/nih-chest-x-ray
BPMLLLoss
false
931
[ "MIT" ]
0
92bfef80f49e7e38ad8a0156be43f10879b3f737
https://github.com/PaulStryck/nih-chest-x-ray/tree/92bfef80f49e7e38ad8a0156be43f10879b3f737
import torch from torch import Tensor class Model(torch.nn.Module): def __init__(self, bias=(1, 1)): super().__init__() self.bias = bias assert len(self.bias) == 2 and all(map(lambda x: isinstance(x, int) and x > 0, bias)), 'bias must be positive integers' def forward(sel...
FirstNet
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class FirstNet(nn.Module): def __init__(self): super(FirstNet, self).__init__() self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size= 3, padding=1, stride=1) self.conv2 = nn.Conv2d(64, 128, 3, pad...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
PacktPublishing/Designing-Models-for-Responsible-AI
FirstNet
false
932
[ "MIT" ]
0
36b60f1e3e9db8b3d2db3ace873dbdee1b076b74
https://github.com/PacktPublishing/Designing-Models-for-Responsible-AI/tree/36b60f1e3e9db8b3d2db3ace873dbdee1b076b74
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(in_channels=1, out_channels=64, kernel_size= 3, padding=1, stride=1) self.conv2 = nn.Conv2d(64, 128, 3, padding=1) def ...
AsymmetricLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class AsymmetricLoss(nn.Module): def __init__(self, gamma_neg=4, gamma_pos=1, clip=0.05, eps=1e-08, disable_torch_grad_focal_loss=False): super(AsymmetricLoss, self).__init__() self.gamma_neg = gamma_neg self.gamma_pos = gamma_pos self.cl...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
Pepijnnn/MasterThesis
AsymmetricLoss
false
933
[ "MIT" ]
0
7ec831f5e55f5f181e0196fa78284e2846ce2e26
https://github.com/Pepijnnn/MasterThesis/tree/7ec831f5e55f5f181e0196fa78284e2846ce2e26
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, gamma_neg=4, gamma_pos=1, clip=0.05, eps=1e-08, disable_torch_grad_focal_loss=False): super().__init__() self.gamma_neg = gamma_neg self.gamma_pos = gamma_pos self.clip = clip self.disabl...