entry_point
stringlengths
1
65
original_triton_code
stringlengths
4.5k
619k
python_code
stringlengths
208
60.9k
triton_code
stringlengths
1.15k
275k
repo_name
stringlengths
7
115
module_name
stringlengths
1
65
synthetic
bool
1 class
uuid
int64
0
18.5k
licenses
listlengths
1
6
stars
int64
0
19.8k
sha
stringlengths
40
40
repo_link
stringlengths
72
180
pytorch_code
stringlengths
200
4.05k
CausalConv1d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class CausalConv1d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=2, dilation=2): super().__init__() self.padding = dilation self.causal_conv = nn.Conv1d(in_channels, out_channels, kernel_size, padding=self.padding, dil...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
Hao-Kailong/DisFeb
CausalConv1d
false
518
[ "MIT" ]
0
2877edd587556e127d6648ee211ed22838c8d015
https://github.com/Hao-Kailong/DisFeb/tree/2877edd587556e127d6648ee211ed22838c8d015
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=2, dilation=2): super().__init__() self.padding = dilation self.causal_conv = nn.Conv1d(in_channels, out_channels, kernel_size, padding=self.padding, dilation=d...
ConvNet2FC
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def spectral_norm(module, init=True, std=1, bound=False): if init: nn.init.normal_(module.weight, 0, std) if hasattr(module, 'bias') and module.bias is not None: module.bias.data.zero_() SpectralNorm.apply(module, 'weight', bound=bound) return module ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
GloryyrolG/normalized-autoencoders
ConvNet2FC
false
519
[ "MIT" ]
0
27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
https://github.com/GloryyrolG/normalized-autoencoders/tree/27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
import torch import torch.nn as nn def spectral_norm(module, init=True, std=1, bound=False): if init: nn.init.normal_(module.weight, 0, std) if hasattr(module, 'bias') and module.bias is not None: module.bias.data.zero_() SpectralNorm.apply(module, 'weight', bound=bound) return module ...
Envelope
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.fx import torch.utils.data class Envelope(torch.nn.Module): def __init__(self, exponent): super(Envelope, self).__init__() self.p = exponent + 1 self.a = -(self.p + 1) * (self.p + 2) / 2 self.b = self.p * (self.p + 2) self.c = -self.p * (self.p + ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.fx import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynam...
HWSelf/pytorch_geometric
Envelope
false
520
[ "MIT" ]
0
c1214de674079b5e39e57c045d0f844b60caf590
https://github.com/HWSelf/pytorch_geometric/tree/c1214de674079b5e39e57c045d0f844b60caf590
import torch import torch.fx import torch.utils.data class Model(torch.nn.Module): def __init__(self, exponent): super().__init__() self.p = exponent + 1 self.a = -(self.p + 1) * (self.p + 2) / 2 self.b = self.p * (self.p + 2) self.c = -self.p * (self.p + 1) / 2 def f...
Attn
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class Attn(torch.nn.Module): """ Attention: feature_dim: dimension of feature embedding method: method to calculate attention, (general, dot, concat) input_dim: dimension of input embedding, default is the same as feature_dim; method dot is only available wh...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
HCDM/XRec
Attn
false
521
[ "MIT" ]
0
dae7d3e1237b8e41913656eb33d81e78c61424ea
https://github.com/HCDM/XRec/tree/dae7d3e1237b8e41913656eb33d81e78c61424ea
import torch from torch import nn class Model(torch.nn.Module): """ Attention: feature_dim: dimension of feature embedding method: method to calculate attention, (general, dot, concat) input_dim: dimension of input embedding, default is the same as feature_dim; method dot is only available w...
Encoder
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class Encoder(nn.Module): """利用卷积 + 最大池化得到句子嵌入""" def __init__(self, max_length, word_embedding_dim=50, pos_embedding_dim =5, hidden_size=230): nn.Module.__init__(self) self.max_length = max_length self.hidden_...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import ...
Hao-Kailong/DisFeb
Encoder
false
522
[ "MIT" ]
0
2877edd587556e127d6648ee211ed22838c8d015
https://github.com/Hao-Kailong/DisFeb/tree/2877edd587556e127d6648ee211ed22838c8d015
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): """利用卷积 + 最大池化得到句子嵌入""" def __init__(self, max_length, word_embedding_dim=50, pos_embedding_dim =5, hidden_size=230): nn.Module.__init__(self) self.max_length = max_length self.hidden_si...
CrossNet
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class CrossNet(nn.Module): """The Cross Network part of Deep&Cross Network model, which leans both low and high degree cross feature. Input shape - 2D tensor with shape: ``(batch_size, units)``. Output shape - 2D tensor with shape: ``(batch_size, u...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_st...
HCDM/XRec
CrossNet
false
523
[ "MIT" ]
0
dae7d3e1237b8e41913656eb33d81e78c61424ea
https://github.com/HCDM/XRec/tree/dae7d3e1237b8e41913656eb33d81e78c61424ea
import torch from torch import nn class Model(nn.Module): """The Cross Network part of Deep&Cross Network model, which leans both low and high degree cross feature. Input shape - 2D tensor with shape: ``(batch_size, units)``. Output shape - 2D tensor with shape: ``(batch_size, unit...
VGGASPP
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class FCReLUDrop(nn.Sequential): def __init__(self, in_ch, out_ch, kernel_size, dilation, padding, layer_idx, branch_idx): super(FCReLUDrop, self).__init__() self.add_module(f'fc{layer_idx}_{branch_idx}', nn.Conv2d(in_ch, out_ch, kernel_size, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
HAL-42/DeepLabV2YQ
VGGASPP
false
524
[ "Apache-2.0" ]
0
96bfcf1055da7adeb4a7c1ed841f6ec29957be59
https://github.com/HAL-42/DeepLabV2YQ/tree/96bfcf1055da7adeb4a7c1ed841f6ec29957be59
import torch from torch import nn class FCReLUDrop(nn.Sequential): def __init__(self, in_ch, out_ch, kernel_size, dilation, padding, layer_idx, branch_idx): super().__init__() self.add_module(f'fc{layer_idx}_{branch_idx}', nn.Conv2d(in_ch, out_ch, kernel_size, stride=1, paddin...
InstanceNorm2D
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class InstanceNorm2D(nn.Module): def __init__(self, num_channels, epsilon=1e-05, momentum=0.9, rescale=True ): super(InstanceNorm2D, self).__init__() self.num_channels = num_channels self.epsilon = epsilon self.momentum = momentum ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
HarmanDotpy/Normalizations-in-Deep-Learning
InstanceNorm2D
false
525
[ "MIT" ]
0
3e1899837fb3ba625f515ef1a995f3573b65456d
https://github.com/HarmanDotpy/Normalizations-in-Deep-Learning/tree/3e1899837fb3ba625f515ef1a995f3573b65456d
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, num_channels, epsilon=1e-05, momentum=0.9, rescale=True ): super().__init__() self.num_channels = num_channels self.epsilon = epsilon self.momentum = momentum self.rescale = rescale ...
NN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.functional as F class NN(nn.Module): def __init__(self, input_size, num_classes): super(NN, self).__init__() self.fc1 = nn.Linear(input_size, 50) self.fc2 = nn.Linear(50, num_classes) def forward(self, x): x = F.relu(self.fc1(...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
HaowenWeiJohn/CV_Project
NN
false
526
[ "MIT" ]
0
8e2414796f60a8c3fe452f3721e4a6ef7edfdb11
https://github.com/HaowenWeiJohn/CV_Project/tree/8e2414796f60a8c3fe452f3721e4a6ef7edfdb11
import torch from torch import nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, input_size, num_classes): super().__init__() self.fc1 = nn.Linear(input_size, 50) self.fc2 = nn.Linear(50, num_classes) def forward(self, x): x = F.relu(self.fc1(x)) ...
LayerNorm2D
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class LayerNorm2D(nn.Module): def __init__(self, num_channels, epsilon=1e-05): super(LayerNorm2D, self).__init__() self.num_channels = num_channels self.epsilon = epsilon self.gamma = nn.Parameter(torch.ones(num_channels)) self.beta = nn....
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
HarmanDotpy/Normalizations-in-Deep-Learning
LayerNorm2D
false
527
[ "MIT" ]
0
3e1899837fb3ba625f515ef1a995f3573b65456d
https://github.com/HarmanDotpy/Normalizations-in-Deep-Learning/tree/3e1899837fb3ba625f515ef1a995f3573b65456d
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, num_channels, epsilon=1e-05): super().__init__() self.num_channels = num_channels self.epsilon = epsilon self.gamma = nn.Parameter(torch.ones(num_channels)) self.beta = nn.Parameter(torch.zeros(n...
GroupNorm2D
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class GroupNorm2D(nn.Module): def __init__(self, num_channels, num_groups=4, epsilon=1e-05): super(GroupNorm2D, self).__init__() self.num_channels = num_channels self.num_groups = num_channels // 4 self.epsilon = epsilon self.gamma = nn.P...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
HarmanDotpy/Normalizations-in-Deep-Learning
GroupNorm2D
false
528
[ "MIT" ]
0
3e1899837fb3ba625f515ef1a995f3573b65456d
https://github.com/HarmanDotpy/Normalizations-in-Deep-Learning/tree/3e1899837fb3ba625f515ef1a995f3573b65456d
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, num_channels, num_groups=4, epsilon=1e-05): super().__init__() self.num_channels = num_channels self.num_groups = num_channels // 4 self.epsilon = epsilon self.gamma = nn.Parameter(torch.ones(num...
InstanceNorm
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class InstanceNorm(nn.Module): def __init__(self, epsilon=1e-08): """ avoid in-place ops. https://discuss.pytorch.org/t/encounter-the-runtimeerror-one-of-the-variables-needed-for-gradient-computation-has-been-modified-by-an-inplace-operation/836/3 """ ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
Holmes-Alan/Photo2Sketch
InstanceNorm
false
529
[ "MIT" ]
0
43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
https://github.com/Holmes-Alan/Photo2Sketch/tree/43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, epsilon=1e-08): """ avoid in-place ops. https://discuss.pytorch.org/t/encounter-the-runtimeerror-one-of-the-variables-needed-for-gradient-computation-has-been-modified-by-an-inplace-operation/836/3 """ super...
TVLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch class TVLoss(torch.nn.Module): def __init__(self): super(TVLoss, self).__init__() def forward(self, x): x.size()[0] h_x = x.size()[2] w_x = x.size()[3] self._tensor_size(x[:, :, 1:, :]) self._tensor_size(x[:, :, :, 1:]) h_tv = torch.pow(x[...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda @triton.j...
Holmes-Alan/Photo2Sketch
TVLoss
false
530
[ "MIT" ]
0
43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
https://github.com/Holmes-Alan/Photo2Sketch/tree/43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
import torch class Model(torch.nn.Module): def __init__(self): super().__init__() def forward(self, x): x.size()[0] h_x = x.size()[2] w_x = x.size()[3] self._tensor_size(x[:, :, 1:, :]) self._tensor_size(x[:, :, :, 1:]) h_tv = torch.pow(x[:, :, 1:, :] ...
Net
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.functional as F class Net(nn.Module): def __init__(self): super(Net, self).__init__() self.conv1 = nn.Conv2d(3, 100, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 = nn.Conv2d(100, 16, 5) self.fc1 = nn.Linear(16 * 5 * 5, 120)...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
Halo1236/Dive-into-DL-PyTorch
Net
false
531
[ "Apache-2.0" ]
0
586b4e9ca77b2121ce5f5bec8b0a893b33f1b574
https://github.com/Halo1236/Dive-into-DL-PyTorch/tree/586b4e9ca77b2121ce5f5bec8b0a893b33f1b574
import torch from torch import nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(3, 100, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 = nn.Conv2d(100, 16, 5) self.fc1 = nn.Linear(16 * 5 * 5, 120) ...
GlobalAvgPool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.nn.functional as F class GlobalAvgPool2d(nn.Module): def __init__(self): super(GlobalAvgPool2d, self).__init__() def forward(self, x): return F.avg_pool2d(x, kernel_size=x.size()[2:]) def get_inputs(): return [torch.rand([4, 4, 4, 4])] d...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_str...
Halo1236/Dive-into-DL-PyTorch
GlobalAvgPool2d
false
532
[ "Apache-2.0" ]
0
586b4e9ca77b2121ce5f5bec8b0a893b33f1b574
https://github.com/Halo1236/Dive-into-DL-PyTorch/tree/586b4e9ca77b2121ce5f5bec8b0a893b33f1b574
import torch from torch import nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() def forward(self, x): return F.avg_pool2d(x, kernel_size=x.size()[2:]) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): retur...
DeConvNet3
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
GloryyrolG/normalized-autoencoders
DeConvNet3
false
533
[ "MIT" ]
0
27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
https://github.com/GloryyrolG/normalized-autoencoders/tree/27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
AvgPoolPad
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from math import * class AvgPoolPad(nn.Module): def __init__(self, stride=2, padding=1): super(AvgPoolPad, self).__init__() self.pad = nn.ZeroPad2d((1, 0, 1, 0)) self.pool = nn.AvgPool2d(3, stride=stride, padding=padding, count_include_pad=Fa...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from math import * assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dyna...
Helicopt/torchreid-preprocess
AvgPoolPad
false
534
[ "MIT" ]
0
2597e502eef079705a5f8a9115a9a1980a9d080d
https://github.com/Helicopt/torchreid-preprocess/tree/2597e502eef079705a5f8a9115a9a1980a9d080d
import torch import torch.nn as nn from math import * class Model(nn.Module): def __init__(self, stride=2, padding=1): super().__init__() self.pad = nn.ZeroPad2d((1, 0, 1, 0)) self.pool = nn.AvgPool2d(3, stride=stride, padding=padding, count_include_pad=False) def forward...
SharpenSoftmax
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class SharpenSoftmax(nn.Module): def __init__(self, tau, dim=0): super().__init__() self.tau = tau self.dim = dim def forward(self, pred): pred = pred / self.tau return pred.log_softmax(self.dim) def get_inputs(): return [torch...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
Hayoung93/UDA
SharpenSoftmax
false
535
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, tau, dim=0): super().__init__() self.tau = tau self.dim = dim def forward(self, pred): pred = pred / self.tau return pred.log_softmax(self.dim) def get_inputs(): return [torch.rand([4,...
GramMatrix
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class GramMatrix(nn.Module): def forward(self, input): b, c, h, w = input.size() f = input.view(b, c, h * w) G = torch.bmm(f, f.transpose(1, 2)) return G.div_(c * h * w) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inp...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
Holmes-Alan/Photo2Sketch
GramMatrix
false
536
[ "MIT" ]
0
43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
https://github.com/Holmes-Alan/Photo2Sketch/tree/43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
import torch import torch.nn as nn class Model(nn.Module): def forward(self, input): b, c, h, w = input.size() f = input.view(b, c, h * w) G = torch.bmm(f, f.transpose(1, 2)) return G.div_(c * h * w) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs()...
HardAttn
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch.nn import functional as F import torch.nn as nn from math import * class HardAttn(nn.Module): """Hard Attention (Sec. 3.1.II)""" def __init__(self, in_channels): super(HardAttn, self).__init__() self.fc = nn.Linear(in_channels, 4 * 2) self.init_params() de...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Helicopt/torchreid-preprocess
HardAttn
false
537
[ "MIT" ]
0
2597e502eef079705a5f8a9115a9a1980a9d080d
https://github.com/Helicopt/torchreid-preprocess/tree/2597e502eef079705a5f8a9115a9a1980a9d080d
import torch from torch.nn import functional as F import torch.nn as nn from math import * class Model(nn.Module): """Hard Attention (Sec. 3.1.II)""" def __init__(self, in_channels): super().__init__() self.fc = nn.Linear(in_channels, 4 * 2) self.init_params() def init_params(sel...
SiamusicLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class SiamusicLoss(nn.Module): def __init__(self, dim=1): super().__init__() self.dim = dim def neg_cos_sim(self, p, z): z = z.detach() p = F.normalize(p, dim=self.dim) z = F.normalize(z, dim=self.dim)...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn import...
HongSungRae/SiamRec
SiamusicLoss
false
538
[ "MIT" ]
0
2ab3b973bc6503eeea66c15c563fdd75b8e5bea1
https://github.com/HongSungRae/SiamRec/tree/2ab3b973bc6503eeea66c15c563fdd75b8e5bea1
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, dim=1): super().__init__() self.dim = dim def neg_cos_sim(self, p, z): z = z.detach() p = F.normalize(p, dim=self.dim) z = F.normalize(z, dim=self.dim) ...
styleLoss_v2
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn def calc_mean_std(feat, eps=1e-05): size = feat.size() assert len(size) == 4 N, C = size[:2] feat_var = feat.view(N, C, -1).var(dim=2) + eps feat_std = feat_var.sqrt().view(N, C, 1, 1) feat_mean = feat.view(N, C, -1).mean(dim=2).view(N, C, 1, 1) return fe...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Holmes-Alan/Photo2Sketch
styleLoss_v2
false
539
[ "MIT" ]
0
43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
https://github.com/Holmes-Alan/Photo2Sketch/tree/43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
import torch import torch.nn as nn def calc_mean_std(feat, eps=1e-05): size = feat.size() assert len(size) == 4 N, C = size[:2] feat_var = feat.view(N, C, -1).var(dim=2) + eps feat_std = feat_var.sqrt().view(N, C, 1, 1) feat_mean = feat.view(N, C, -1).mean(dim=2).view(N, C, 1, 1) return fe...
InnerProductDecoder
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.modules.loss class InnerProductDecoder(nn.Module): """Decoder for using inner product for prediction.""" def __init__(self, dropout, act=torch.sigmoid): super(InnerProductDecoder, self).__init__() self.dropout ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.modules.loss assert_size_stride = torch._C...
HongyiZhu/EHI
InnerProductDecoder
false
540
[ "MIT" ]
0
9fbbc6046546dd7fc6de5d831b4c941bc4404e02
https://github.com/HongyiZhu/EHI/tree/9fbbc6046546dd7fc6de5d831b4c941bc4404e02
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.modules.loss class Model(nn.Module): """Decoder for using inner product for prediction.""" def __init__(self, dropout, act=torch.sigmoid): super().__init__() self.dropout = dropout self.act = act d...
AddReadout
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class AddReadout(nn.Module): """Handles readout operation when `readout` parameter is `add`. Removes `cls_token` or `readout_token` from tensor and adds it to the rest of tensor""" def __init__(self, start_index=1): super(AddReadout, self).__init__() self.s...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
HrithikNambiar/vformer
AddReadout
false
541
[ "MIT" ]
0
5bd902a45e5cae70ab001ca6c217f12f923561f1
https://github.com/HrithikNambiar/vformer/tree/5bd902a45e5cae70ab001ca6c217f12f923561f1
import torch import torch.nn as nn class Model(nn.Module): """Handles readout operation when `readout` parameter is `add`. Removes `cls_token` or `readout_token` from tensor and adds it to the rest of tensor""" def __init__(self, start_index=1): super().__init__() self.start_index = start_in...
MaxPoolPad
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from math import * class MaxPoolPad(nn.Module): def __init__(self): super(MaxPoolPad, self).__init__() self.pad = nn.ZeroPad2d((1, 0, 1, 0)) self.pool = nn.MaxPool2d(3, stride=2, padding=1) def forward(self, x): x = self.pad(x) x = s...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn from math import * assert_size_stride = torch._C._dynamo.guards.ass...
Helicopt/torchreid-preprocess
MaxPoolPad
false
542
[ "MIT" ]
0
2597e502eef079705a5f8a9115a9a1980a9d080d
https://github.com/Helicopt/torchreid-preprocess/tree/2597e502eef079705a5f8a9115a9a1980a9d080d
import torch import torch.nn as nn from math import * class Model(nn.Module): def __init__(self): super().__init__() self.pad = nn.ZeroPad2d((1, 0, 1, 0)) self.pool = nn.MaxPool2d(3, stride=2, padding=1) def forward(self, x): x = self.pad(x) x = self.pool(x) x...
DeConvNet64
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
GloryyrolG/normalized-autoencoders
DeConvNet64
false
543
[ "MIT" ]
0
27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
https://github.com/GloryyrolG/normalized-autoencoders/tree/27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
RoundPass
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch as t import torch.utils.data class RoundPass(t.nn.Module): def forward(self, x): y = x.round() y_grad = x return (y - y_grad).detach() + y_grad def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], {}]
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch as t import torch.utils.data assert_size_stride = torch._C._dynamo...
HumberMe/lsq-net
RoundPass
false
544
[ "MIT" ]
0
7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
https://github.com/HumberMe/lsq-net/tree/7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
import torch import torch as t import torch.utils.data class Model(t.nn.Module): def forward(self, x): y = x.round() y_grad = x return (y - y_grad).detach() + y_grad def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []
GraphConvolution
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class GraphConvolution(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/1609.02907 """ def __init__(...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch.nn import Module i...
HongyiZhu/EHI
GraphConvolution
false
545
[ "MIT" ]
0
9fbbc6046546dd7fc6de5d831b4c941bc4404e02
https://github.com/HongyiZhu/EHI/tree/9fbbc6046546dd7fc6de5d831b4c941bc4404e02
from torch.nn import Module import torch import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class Model(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/1609.02907 """ def __init__(self, in_fe...
HardMish
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.parallel def hard_mish(x, inplace: 'bool'=False): """ Hard Mish Experimental, based on notes by Mish author Diganta Misra at https://github.com/digantamisra98/H-Mish/blob/0da20d4bc58e696b6803f2523c58d3c8a82782d0/README.md """ if inplace: ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import torch.nn.parallel assert_size_stride = torch._C._dynamo.guar...
HotaekHan/detr_pytorch
HardMish
false
546
[ "MIT" ]
0
730e02db0ac8910ef782234a3990587771ad67f9
https://github.com/HotaekHan/detr_pytorch/tree/730e02db0ac8910ef782234a3990587771ad67f9
import torch import torch.nn as nn import torch.nn.parallel def hard_mish(x, inplace: 'bool'=False): """ Hard Mish Experimental, based on notes by Mish author Diganta Misra at https://github.com/digantamisra98/H-Mish/blob/0da20d4bc58e696b6803f2523c58d3c8a82782d0/README.md """ if inplace: ...
GlobalAvgPool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.parallel class GlobalAvgPool2d(nn.Module): def __init__(self): """Global average pooling over the input's spatial dimensions""" super(GlobalAvgPool2d, self).__init__() def forward(self, inputs): return nn.functional.adaptive_avg_pool...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.parallel assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C...
HotaekHan/detr_pytorch
GlobalAvgPool2d
false
547
[ "MIT" ]
0
730e02db0ac8910ef782234a3990587771ad67f9
https://github.com/HotaekHan/detr_pytorch/tree/730e02db0ac8910ef782234a3990587771ad67f9
import torch import torch.nn as nn import torch.nn.parallel class Model(nn.Module): def __init__(self): """Global average pooling over the input's spatial dimensions""" super().__init__() def forward(self, inputs): return nn.functional.adaptive_avg_pool2d(inputs, 1).view(inputs. ...
Selection
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Selection(nn.Module): """ Selection neurons to sample from a latent representation for a decoder agent. An abstract representation :math:`l_i` is disturbed by a value :math:`r_i` sampled from a normal standard distribution which is scaled by the selection neur...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torch.nn as nn assert_size_stride = torch._C._dynamo.gu...
HendrikPN/intervention-based-autoencoder
Selection
false
548
[ "Apache-2.0" ]
0
90018d8ea264681cc9b9b55ba9e531e36275136f
https://github.com/HendrikPN/intervention-based-autoencoder/tree/90018d8ea264681cc9b9b55ba9e531e36275136f
import torch import torch.nn as nn class Model(nn.Module): """ Selection neurons to sample from a latent representation for a decoder agent. An abstract representation :math:`l_i` is disturbed by a value :math:`r_i` sampled from a normal standard distribution which is scaled by the selection neuron :...
Pad_Pool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn class Pad_Pool2d(nn.Module): """ Implements a padding layer in front of pool1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, left=0, right=1, top=0, bottom=1, value=0...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_str...
Hullimulli/EEGEyeNet
Pad_Pool2d
false
549
[ "MIT" ]
0
677a791b39800f44dc254553b16ee2f92e62c423
https://github.com/Hullimulli/EEGEyeNet/tree/677a791b39800f44dc254553b16ee2f92e62c423
import torch from torch import nn class Model(nn.Module): """ Implements a padding layer in front of pool1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, left=0, right=1, top=0, bottom=1, value=0): ...
GCNModelVAE
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class GraphConvolution(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/1609.02907 ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch.nn import Module i...
HongyiZhu/EHI
GCNModelVAE
false
550
[ "MIT" ]
0
9fbbc6046546dd7fc6de5d831b4c941bc4404e02
https://github.com/HongyiZhu/EHI/tree/9fbbc6046546dd7fc6de5d831b4c941bc4404e02
from torch.nn import Module import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class GraphConvolution(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/1609.02907 ...
Pad_Conv2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import math import torch from torch import nn class Pad_Conv2d(nn.Module): """ Implements a padding layer in front of conv2d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x Input: kernel as a tuple (kx, ky) Output: Pad...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import math from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guard...
Hullimulli/EEGEyeNet
Pad_Conv2d
false
551
[ "MIT" ]
0
677a791b39800f44dc254553b16ee2f92e62c423
https://github.com/Hullimulli/EEGEyeNet/tree/677a791b39800f44dc254553b16ee2f92e62c423
import math import torch from torch import nn class Model(nn.Module): """ Implements a padding layer in front of conv2d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x Input: kernel as a tuple (kx, ky) Output: Padded t...
PatchEmbedding
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def pair(t): """ Parameters ---------- t: tuple[int] or int """ return t if isinstance(t, tuple) else (t, t) class PatchEmbedding(nn.Module): """ Parameters ---------- img_size: int Image Size patch_size: int Patch Size ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
HrithikNambiar/vformer
PatchEmbedding
false
552
[ "MIT" ]
0
5bd902a45e5cae70ab001ca6c217f12f923561f1
https://github.com/HrithikNambiar/vformer/tree/5bd902a45e5cae70ab001ca6c217f12f923561f1
import torch import torch.nn as nn def pair(t): """ Parameters ---------- t: tuple[int] or int """ return t if isinstance(t, tuple) else (t, t) class Model(nn.Module): """ Parameters ---------- img_size: int Image Size patch_size: int Patch Size in_ch...
SEModule
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.parallel class FastGlobalAvgPool2d(nn.Module): def __init__(self, flatten=False): super(FastGlobalAvgPool2d, self).__init__() self.flatten = flatten def forward(self, x): if self.flatten: in_size = x.size() re...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import ...
HotaekHan/detr_pytorch
SEModule
false
553
[ "MIT" ]
0
730e02db0ac8910ef782234a3990587771ad67f9
https://github.com/HotaekHan/detr_pytorch/tree/730e02db0ac8910ef782234a3990587771ad67f9
import torch import torch.nn as nn import torch.nn.parallel class FastGlobalAvgPool2d(nn.Module): def __init__(self, flatten=False): super().__init__() self.flatten = flatten def forward(self, x): if self.flatten: in_size = x.size() return x.view((in_size[0], ...
ConvNet64
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
GloryyrolG/normalized-autoencoders
ConvNet64
false
554
[ "MIT" ]
0
27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
https://github.com/GloryyrolG/normalized-autoencoders/tree/27ccb74bb725768f9ba9ea6fa03a7a40867eebb1
import torch import torch.nn as nn def get_activation(s_act): if s_act == 'relu': return nn.ReLU(inplace=True) elif s_act == 'sigmoid': return nn.Sigmoid() elif s_act == 'softplus': return nn.Softplus() elif s_act == 'linear': return None elif s_act == 'tanh': ...
FocalLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss tensor. """ ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
HumberMe/mmclassification
FocalLoss
false
555
[ "Apache-2.0" ]
0
68f1542068d3af4db932c97e6a728181432fff0c
https://github.com/HumberMe/mmclassification/tree/68f1542068d3af4db932c97e6a728181432fff0c
import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss tensor. """ ...
LsqQuan
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch as t import torch.utils.data class GradScale(t.nn.Module): def forward(self, x, scale): y = x y_grad = x * scale return (y - y_grad).detach() + y_grad class RoundPass(t.nn.Module): def forward(self, x): y = x.round() y_grad = x retu...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import torch as t import tor...
HumberMe/lsq-net
LsqQuan
false
556
[ "MIT" ]
0
7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
https://github.com/HumberMe/lsq-net/tree/7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
import torch import torch as t import torch.utils.data class GradScale(t.nn.Module): def forward(self, x, scale): y = x y_grad = x * scale return (y - y_grad).detach() + y_grad class RoundPass(t.nn.Module): def forward(self, x): y = x.round() y_grad = x retu...
ChamferLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed from typing import * class ChamferLoss(nn.Module): def __init__(self): super(ChamferLoss, self).__init__() self.use_cuda = torch.cuda.is_available() def b...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn import t...
DeVriesMatt/pointMLP-pytorch
ChamferLoss
false
557
[ "Apache-2.0" ]
0
e9c09a2038551e83b072353f3fd7e3294463e892
https://github.com/DeVriesMatt/pointMLP-pytorch/tree/e9c09a2038551e83b072353f3fd7e3294463e892
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed from typing import * class Model(nn.Module): def __init__(self): super().__init__() self.use_cuda = torch.cuda.is_available() def batch_pairwise_dist(self...
Actor
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn.functional as F import torch.nn as nn def hidden_init(layer): fan_in = layer.weight.data.size()[0] lim = 1.0 / np.sqrt(fan_in) return -lim, lim class Actor(nn.Module): """Actor (Policy) Model.""" def __init__(self, state_size, action_size, seed, f...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
HuXiao-THU/Crane-Group-Control
Actor
false
558
[ "MIT" ]
0
ea71bc9b1e3957fd755312ceb52bda1be8244f5a
https://github.com/HuXiao-THU/Crane-Group-Control/tree/ea71bc9b1e3957fd755312ceb52bda1be8244f5a
import torch import numpy as np import torch.nn.functional as F import torch.nn as nn def hidden_init(layer): fan_in = layer.weight.data.size()[0] lim = 1.0 / np.sqrt(fan_in) return -lim, lim class Model(nn.Module): """Actor (Policy) Model.""" def __init__(self, state_size, action_size, seed, f...
GCNAutoencoder
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import math import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class GraphConvolutionDecoder(Module): """ Simple GCN layer, similar to https://arxiv.org/...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch.nn import Module i...
HongyiZhu/EHI
GCNAutoencoder
false
559
[ "MIT" ]
0
9fbbc6046546dd7fc6de5d831b4c941bc4404e02
https://github.com/HongyiZhu/EHI/tree/9fbbc6046546dd7fc6de5d831b4c941bc4404e02
from torch.nn import Module import math import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.parameter import Parameter from torch.nn.modules.module import Module import torch.nn.modules.loss class GraphConvolutionDecoder(Module): """ Simple GCN layer, similar to https://arxiv.org/...
Pad_Pool
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn class Pad_Pool(nn.Module): """ Implements a padding layer in front of pool1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, left=0, right=1, value=0): super()....
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_str...
Hullimulli/EEGEyeNet
Pad_Pool
false
560
[ "MIT" ]
0
677a791b39800f44dc254553b16ee2f92e62c423
https://github.com/Hullimulli/EEGEyeNet/tree/677a791b39800f44dc254553b16ee2f92e62c423
import torch from torch import nn class Model(nn.Module): """ Implements a padding layer in front of pool1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, left=0, right=1, value=0): super().__i...
LayerNorm2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class LayerNorm2d(nn.LayerNorm): """LayerNorm on channels for 2d images. Args: num_channels (int): The number of channels of the input tensor. eps (float): a value added to the denominator for numerical stability. ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
HumberMe/mmclassification
LayerNorm2d
false
561
[ "Apache-2.0" ]
0
68f1542068d3af4db932c97e6a728181432fff0c
https://github.com/HumberMe/mmclassification/tree/68f1542068d3af4db932c97e6a728181432fff0c
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.LayerNorm): """LayerNorm on channels for 2d images. Args: num_channels (int): The number of channels of the input tensor. eps (float): a value added to the denominator for numerical stability. Defaul...
GeneralizedMeanPooling
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import Tensor import torch.nn as nn from torch.functional import Tensor import torch.nn.functional as F from torch import Tensor from torch.nn.parameter import Parameter def gem(x: 'Tensor', p: 'Parameter', eps: 'float'=1e-06, clamp=True) ->Tensor: if clamp: x = x.clamp(min=eps) ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math from torch ...
HumberMe/mmclassification
GeneralizedMeanPooling
false
562
[ "Apache-2.0" ]
0
68f1542068d3af4db932c97e6a728181432fff0c
https://github.com/HumberMe/mmclassification/tree/68f1542068d3af4db932c97e6a728181432fff0c
import torch from torch import Tensor import torch.nn as nn from torch.functional import Tensor import torch.nn.functional as F from torch import Tensor from torch.nn.parameter import Parameter def gem(x: 'Tensor', p: 'Parameter', eps: 'float'=1e-06, clamp=True) ->Tensor: if clamp: x = x.clamp(min=eps) ...
LN
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class LN(nn.Module): def forward(self, x): return F.layer_norm(x, x.size()[1:], weight=None, bias=None, eps=1e-05) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn import torch.utils.data assert_size_stride = torch._C._dy...
ID56/OrigamiNet
LN
false
563
[ "Apache-2.0" ]
0
a71ec4984e3d5da7d635d68260026b749ec44fa9
https://github.com/ID56/OrigamiNet/tree/a71ec4984e3d5da7d635d68260026b749ec44fa9
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class Model(nn.Module): def forward(self, x): return F.layer_norm(x, x.size()[1:], weight=None, bias=None, eps=1e-05) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []...
ResidualBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class ResidualBlock(nn.Module): def __init__(self, input_channel, output_channel, upsample=True): super(ResidualBlock, self).__init__() self.conv1 = nn.Conv2d(input_channel, output_channel, kernel_size=3, padding=0) ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Holmes-Alan/Photo2Sketch
ResidualBlock
false
564
[ "MIT" ]
0
43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
https://github.com/Holmes-Alan/Photo2Sketch/tree/43a0ca6bb8a8e645b35a2ab23d11ed5efe117e09
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, input_channel, output_channel, upsample=True): super().__init__() self.conv1 = nn.Conv2d(input_channel, output_channel, kernel_size=3, padding=0) self.conv2 = nn.Conv2...
GradScale
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch as t import torch.utils.data class GradScale(t.nn.Module): def forward(self, x, scale): y = x y_grad = x * scale return (y - y_grad).detach() + y_grad def get_inputs(): return [torch.rand([4, 4, 4, 4]), torch.rand([4, 4, 4, 4])] def get_init_inputs(): ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch as t import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dyn...
HumberMe/lsq-net
GradScale
false
565
[ "MIT" ]
0
7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
https://github.com/HumberMe/lsq-net/tree/7dcd75bff4aa7ff2d9c8a7902198fe411a38eb4c
import torch import torch as t import torch.utils.data class Model(t.nn.Module): def forward(self, x, scale): y = x y_grad = x * scale return (y - y_grad).detach() + y_grad def get_inputs(): return [torch.rand([4, 4, 4, 4]), torch.rand([4, 4, 4, 4])] def get_init_inputs(): ret...
Pad_Conv
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import math import torch from torch import nn class Pad_Conv(nn.Module): """ Implements a padding layer in front of conv1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, kernel_size, value=0): ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import math from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guard...
Hullimulli/EEGEyeNet
Pad_Conv
false
566
[ "MIT" ]
0
677a791b39800f44dc254553b16ee2f92e62c423
https://github.com/Hullimulli/EEGEyeNet/tree/677a791b39800f44dc254553b16ee2f92e62c423
import math import torch from torch import nn class Model(nn.Module): """ Implements a padding layer in front of conv1d layers used in our architectures to achieve padding=same output shape Pads 0 to the left and 1 to the right side of x """ def __init__(self, kernel_size, value=0): sup...
ZeroConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from torch.nn import functional as F class ZeroConv2d(nn.Module): def __init__(self, in_channel, out_channel, padding=1): super().__init__() self.conv = nn.Conv2d(in_channel, out_channel, 3, padding=0) self.conv.weight.data.zero_() self.conv.bias....
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math from torch im...
HugoSenetaire/glow-pytorch
ZeroConv2d
false
567
[ "MIT" ]
0
7f11be87cac9770df63867910c34738dedee6f56
https://github.com/HugoSenetaire/glow-pytorch/tree/7f11be87cac9770df63867910c34738dedee6f56
import torch from torch import nn from torch.nn import functional as F class Model(nn.Module): def __init__(self, in_channel, out_channel, padding=1): super().__init__() self.conv = nn.Conv2d(in_channel, out_channel, 3, padding=0) self.conv.weight.data.zero_() self.conv.bias.data....
ArctanLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn from abc import ABCMeta from abc import abstractmethod class Layer(nn.Module, metaclass=ABCMeta): def __init__(self): super(Layer, self).__init__() @abstractmethod def forward(self, x): """ >>> do forward pass with a given input ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn import torch.nn from abc import ABCMeta from abc import a...
Isaac-Li-cn/certify_robustness
ArctanLayer
false
568
[ "BSD-3-Clause" ]
0
f904dc923afc6354e406c57a1c923d13fc39d315
https://github.com/Isaac-Li-cn/certify_robustness/tree/f904dc923afc6354e406c57a1c923d13fc39d315
import torch import torch.nn as nn import torch.nn from abc import ABCMeta from abc import abstractmethod class Layer(nn.Module, metaclass=ABCMeta): def __init__(self): super().__init__() @abstractmethod def forward(self, x): """ >>> do forward pass with a given input """...
InstanceNormLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch from torch import nn class InstanceNormLayer(nn.Module): """Implements instance normalization layer.""" def __init__(self, epsilon=1e-08): super().__init__() self.eps = epsilon def forward(self, x): if len(x.shape) != 4: ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.utils.data import torch from torch import nn assert_size_stride = ...
IVRL/BIGPrior
InstanceNormLayer
false
569
[ "MIT" ]
0
6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
https://github.com/IVRL/BIGPrior/tree/6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
import torch import torch.utils.data import torch from torch import nn class Model(nn.Module): """Implements instance normalization layer.""" def __init__(self, epsilon=1e-08): super().__init__() self.eps = epsilon def forward(self, x): if len(x.shape) != 4: raise Val...
Critic
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn.functional as F import torch.nn as nn def hidden_init(layer): fan_in = layer.weight.data.size()[0] lim = 1.0 / np.sqrt(fan_in) return -lim, lim class Critic(nn.Module): """Critic (Value) Model.""" def __init__(self, state_size, action_size, seed, ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import numpy as np import torch.nn as nn assert_size_stride = torch._C._dynamo.g...
HuXiao-THU/Crane-Group-Control
Critic
false
570
[ "MIT" ]
0
ea71bc9b1e3957fd755312ceb52bda1be8244f5a
https://github.com/HuXiao-THU/Crane-Group-Control/tree/ea71bc9b1e3957fd755312ceb52bda1be8244f5a
import torch import numpy as np import torch.nn.functional as F import torch.nn as nn def hidden_init(layer): fan_in = layer.weight.data.size()[0] lim = 1.0 / np.sqrt(fan_in) return -lim, lim class Model(nn.Module): """Critic (Value) Model.""" def __init__(self, state_size, action_size, seed, f...
AsymmetricLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss tensor. """ ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
HumberMe/mmclassification
AsymmetricLoss
false
571
[ "Apache-2.0" ]
0
68f1542068d3af4db932c97e6a728181432fff0c
https://github.com/HumberMe/mmclassification/tree/68f1542068d3af4db932c97e6a728181432fff0c
import torch import torch.nn as nn import torch.nn.functional as F def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Return: Tensor: Reduced loss tensor. """ ...
ResolutionScalingLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch from torch import nn import torch.nn.functional as F class ResolutionScalingLayer(nn.Module): """Implements the resolution scaling layer. Basically, this layer can be used to upsample feature maps from spatial domain with nearest neighbor interpolation. """...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.utils.data import torch from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda...
IVRL/BIGPrior
ResolutionScalingLayer
false
572
[ "MIT" ]
0
6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
https://github.com/IVRL/BIGPrior/tree/6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
import torch import torch.utils.data import torch from torch import nn import torch.nn.functional as F class Model(nn.Module): """Implements the resolution scaling layer. Basically, this layer can be used to upsample feature maps from spatial domain with nearest neighbor interpolation. """ def __init_...
ReLULayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn import torch.nn.functional as F from abc import ABCMeta from abc import abstractmethod class Layer(nn.Module, metaclass=ABCMeta): def __init__(self): super(Layer, self).__init__() @abstractmethod def forward(self, x): """ >>> do ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import torch.nn import torch.nn.functional as F from abc import ABC...
Isaac-Li-cn/certify_robustness
ReLULayer
false
573
[ "BSD-3-Clause" ]
0
f904dc923afc6354e406c57a1c923d13fc39d315
https://github.com/Isaac-Li-cn/certify_robustness/tree/f904dc923afc6354e406c57a1c923d13fc39d315
import torch import torch.nn as nn import torch.nn import torch.nn.functional as F from abc import ABCMeta from abc import abstractmethod class Layer(nn.Module, metaclass=ABCMeta): def __init__(self): super().__init__() @abstractmethod def forward(self, x): """ >>> do forward pas...
EqualConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from math import sqrt assert_size_stride = torch._C._dynam...
IW276/IW276SS20-P3
EqualConv2d
false
574
[ "MIT" ]
0
7970bd332cc021cf1879f326c444eff3cf8593a1
https://github.com/IW276/IW276SS20-P3/tree/7970bd332cc021cf1879f326c444eff3cf8593a1
import torch import torch.nn as nn from math import sqrt def equal_lr(module, name='weight'): EqualLR.apply(module, name) return module class EqualLR: def __init__(self, name): self.name = name def compute_weight(self, module): weight = getattr(module, self.name + '_orig') ...
ResidualBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data import torch.nn.functional as F class ResidualBlock(nn.Module): def __init__(self, channel_num, dilation=1, group=1): super(ResidualBlock, self).__init__() self.conv1 = nn.Conv2d(channel_num, channel_num, 3, 1, pa...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Hwihuni/Deep-Model-Watermarking
ResidualBlock
false
575
[ "MIT" ]
0
73ea2286ace0aac3d55f6056da38ea2bc38ed00d
https://github.com/Hwihuni/Deep-Model-Watermarking/tree/73ea2286ace0aac3d55f6056da38ea2bc38ed00d
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data import torch.nn.functional as F class Model(nn.Module): def __init__(self, channel_num, dilation=1, group=1): super().__init__() self.conv1 = nn.Conv2d(channel_num, channel_num, 3, 1, padding= dilation...
BiasLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class BiasLayer(nn.Module): def __init__(self, channels, skip_dims=2): super().__init__() self.bias = nn.Parameter(torch.zeros(channels, *([1] * skip_dims))) def forward(self, net): return net + self.bias def extra_repr(self): return f'...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
JGU-VC/activation-pattern-analysis
BiasLayer
false
576
[ "MIT" ]
0
14da42ad541ee4faf35d360a6e871fd44decd33d
https://github.com/JGU-VC/activation-pattern-analysis/tree/14da42ad541ee4faf35d360a6e871fd44decd33d
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, channels, skip_dims=2): super().__init__() self.bias = nn.Parameter(torch.zeros(channels, *([1] * skip_dims))) def forward(self, net): return net + self.bias def extra_repr(self): return f'shap...
InvConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from torch.nn import functional as F class InvConv2d(nn.Module): def __init__(self, in_channel): super().__init__() weight = torch.randn(in_channel, in_channel) q, _ = torch.qr(weight) weight = q.unsqueeze(2).unsqueeze(3) self.weight = nn....
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn from torch.nn import functional as F assert_size_stride = t...
HugoSenetaire/glow-pytorch
InvConv2d
false
577
[ "MIT" ]
0
7f11be87cac9770df63867910c34738dedee6f56
https://github.com/HugoSenetaire/glow-pytorch/tree/7f11be87cac9770df63867910c34738dedee6f56
import torch from torch import nn from torch.nn import functional as F class Model(nn.Module): def __init__(self, in_channel): super().__init__() weight = torch.randn(in_channel, in_channel) q, _ = torch.qr(weight) weight = q.unsqueeze(2).unsqueeze(3) self.weight = nn.Para...
PixelNormLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch from torch import nn class PixelNormLayer(nn.Module): """Implements pixel-wise feature vector normalization layer.""" def __init__(self, epsilon=1e-08): super().__init__() self.eps = epsilon def forward(self, x): return x / torch....
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.utils.data import torch from torch import nn assert_size_stride = ...
IVRL/BIGPrior
PixelNormLayer
false
578
[ "MIT" ]
0
6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
https://github.com/IVRL/BIGPrior/tree/6bf3b18fcbbd3c58bad7a792a8d28b017abb2411
import torch import torch.utils.data import torch from torch import nn class Model(nn.Module): """Implements pixel-wise feature vector normalization layer.""" def __init__(self, epsilon=1e-08): super().__init__() self.eps = epsilon def forward(self, x): return x / torch.sqrt(torc...
PixelNorm
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils class PixelNorm(nn.Module): def __init__(self, epsilon=1e-08): """ @notice: avoid in-place ops. https://discuss.pytorch.org/t/encounter-the-runtimeerror-one-of-the-variables-needed-for-gradient-com...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import nn import torch.nn.parallel import torch.utils.data import to...
IdanAzuri/MixMatch-pytorch
PixelNorm
false
579
[ "MIT" ]
0
b8de2bc30c09e1256b92e0394403487fc4f90135
https://github.com/IdanAzuri/MixMatch-pytorch/tree/b8de2bc30c09e1256b92e0394403487fc4f90135
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils class Model(nn.Module): def __init__(self, epsilon=1e-08): """ @notice: avoid in-place ops. https://discuss.pytorch.org/t/encounter-the-runtimeerror-one-of-the-variables-needed-for-gradient-computa...
BinaryCrossEntropy2D
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torch.nn.modules.loss import _WeightedLoss class BinaryCrossEntropy2D(_WeightedLoss): """ Standard pytorch weighted nn.CrossEntropyLoss """ def __init__(self): super(BinaryCrossEntropy2D, self).__init__() self.nll_loss = nn.BCELoss(reduction='no...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
JHalili/KidneySegQuicknat
BinaryCrossEntropy2D
false
580
[ "MIT" ]
0
4ddc30f2cf935045bf6482a73a3e86e2d8da3696
https://github.com/JHalili/KidneySegQuicknat/tree/4ddc30f2cf935045bf6482a73a3e86e2d8da3696
import torch import torch.nn as nn from torch.nn.modules.loss import _WeightedLoss class Model(_WeightedLoss): """ Standard pytorch weighted nn.CrossEntropyLoss """ def __init__(self): super().__init__() self.nll_loss = nn.BCELoss(reduction='none') def forward(self, inputs, targe...
IIDIsotropicGaussianUVLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import math import torch import torch.utils.data from torch import nn import torch.nn.functional as F class IIDIsotropicGaussianUVLoss(nn.Module): """ Loss for the case of iid residuals with isotropic covariance: $Sigma_i = sigma_i^2 I$ The loss (negative log likelihood) is then: $1/2 sum_{i=1}^n ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import math...
JHMeusener/detectron2-ResNeSt
IIDIsotropicGaussianUVLoss
false
581
[ "Apache-2.0" ]
0
6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
https://github.com/JHMeusener/detectron2-ResNeSt/tree/6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
import math import torch import torch.utils.data from torch import nn import torch.nn.functional as F class Model(nn.Module): """ Loss for the case of iid residuals with isotropic covariance: $Sigma_i = sigma_i^2 I$ The loss (negative log likelihood) is then: $1/2 sum_{i=1}^n (log(2 pi) + 2 log si...
IndepAnisotropicGaussianUVLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import math import torch import torch.utils.data from torch import nn import torch.nn.functional as F class IndepAnisotropicGaussianUVLoss(nn.Module): """ Loss for the case of independent residuals with anisotropic covariances: $Sigma_i = sigma_i^2 I + r_i r_i^T$ The loss (negative log likelihood) is ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import math...
JHMeusener/detectron2-ResNeSt
IndepAnisotropicGaussianUVLoss
false
582
[ "Apache-2.0" ]
0
6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
https://github.com/JHMeusener/detectron2-ResNeSt/tree/6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
import math import torch import torch.utils.data from torch import nn import torch.nn.functional as F class Model(nn.Module): """ Loss for the case of independent residuals with anisotropic covariances: $Sigma_i = sigma_i^2 I + r_i r_i^T$ The loss (negative log likelihood) is then: $1/2 sum_{i=1}^...
L0Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import * from torch import nn class L0Loss(nn.Module): """L0loss from "Noise2Noise: Learning Image Restoration without Clean Data" <https://arxiv.org/pdf/1803.04189>`_ paper. """ def __init__(self, gamma=2, eps=1e-08): super(L0Loss, self).__init__() self.gamma = g...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from typing import * f...
JacobARose/image-utils
L0Loss
false
583
[ "MIT" ]
0
aa0e005c0b4df5198d188b074f4e21f8d8f97962
https://github.com/JacobARose/image-utils/tree/aa0e005c0b4df5198d188b074f4e21f8d8f97962
import torch from typing import * from torch import nn class Model(nn.Module): """L0loss from "Noise2Noise: Learning Image Restoration without Clean Data" <https://arxiv.org/pdf/1803.04189>`_ paper. """ def __init__(self, gamma=2, eps=1e-08): super().__init__() self.gamma = gamma ...
OHEM_CrossEntroy_Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn class OHEM_CrossEntroy_Loss(nn.Module): def __init__(self, threshold, keep_num): super(OHEM_CrossEntroy_Loss, self).__init__() self.threshold = threshold self.keep_num = keep_num self.loss_function = nn.CrossEntropyLoss(reduction='none') def ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from torch import nn a...
HaowenWeiJohn/CV_Project
OHEM_CrossEntroy_Loss
false
584
[ "MIT" ]
0
8e2414796f60a8c3fe452f3721e4a6ef7edfdb11
https://github.com/HaowenWeiJohn/CV_Project/tree/8e2414796f60a8c3fe452f3721e4a6ef7edfdb11
import torch from torch import nn class Model(nn.Module): def __init__(self, threshold, keep_num): super().__init__() self.threshold = threshold self.keep_num = keep_num self.loss_function = nn.CrossEntropyLoss(reduction='none') def forward(self, output, target): loss...
TwoLayerFCBodyWithAction
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F def layer_init(layer, w_scale=1.0): nn.init.orthogonal_(layer.weight.data) layer.weight.data.mul_(w_scale) nn.init.constant_(layer.bias.data, 0) return layer class TwoLayerFCBodyWithAction(nn.Module): def __init__(self, state_di...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import ...
Fieps1/p3-tennis
TwoLayerFCBodyWithAction
false
585
[ "MIT" ]
0
29f3dab5810d7cd7f84120416a615956d266c256
https://github.com/Fieps1/p3-tennis/tree/29f3dab5810d7cd7f84120416a615956d266c256
import torch import torch.nn as nn import torch.nn.functional as F def layer_init(layer, w_scale=1.0): nn.init.orthogonal_(layer.weight.data) layer.weight.data.mul_(w_scale) nn.init.constant_(layer.bias.data, 0) return layer class Model(nn.Module): def __init__(self, state_dim, action_dim, hidd...
Conv2dSame
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn class Conv2dSame(torch.nn.Module): """2D convolution that pads to keep spatial dimensions equal. Cannot deal with stride. Only quadratic kernels (=scalar kernel_size). """ def __init__(self, in_channels, out_channels, kernel_size, bias=True, p...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math from torch im...
Jack12xl/scene-representation-networks
Conv2dSame
false
586
[ "MIT" ]
0
2691b23c956cf188a1fe4c84a888b19871cac8f4
https://github.com/Jack12xl/scene-representation-networks/tree/2691b23c956cf188a1fe4c84a888b19871cac8f4
import torch from torch import nn import torch.nn class Model(torch.nn.Module): """2D convolution that pads to keep spatial dimensions equal. Cannot deal with stride. Only quadratic kernels (=scalar kernel_size). """ def __init__(self, in_channels, out_channels, kernel_size, bias=True, paddin...
CharbonnierLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class CharbonnierLoss(nn.Module): """Charbonnier Loss (L1)""" def __init__(self, eps=1e-06): super(CharbonnierLoss, self).__init__() self.eps = eps def forward(self, x, y): diff = x - y loss = torch.sum(torch.sqrt...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn import...
JaguAroo/SRResCGAN
CharbonnierLoss
false
587
[ "MIT" ]
0
9aac612aff631f7fb9142e0a36de9559cfc1a62d
https://github.com/JaguAroo/SRResCGAN/tree/9aac612aff631f7fb9142e0a36de9559cfc1a62d
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): """Charbonnier Loss (L1)""" def __init__(self, eps=1e-06): super().__init__() self.eps = eps def forward(self, x, y): diff = x - y loss = torch.sum(torch.sqrt(diff * diff + self.eps)) ...
Classifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils class Classifier(nn.Module): def __init__(self, num_classes, dim=128): super(Classifier, self).__init__() self.num_classes = num_classes self.dim = dim self.fc1 = nn.Linear(self.d...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn import torch.nn.parallel import torch.utils.data import tor...
IdanAzuri/MixMatch-pytorch
Classifier
false
588
[ "MIT" ]
0
b8de2bc30c09e1256b92e0394403487fc4f90135
https://github.com/IdanAzuri/MixMatch-pytorch/tree/b8de2bc30c09e1256b92e0394403487fc4f90135
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils class Model(nn.Module): def __init__(self, num_classes, dim=128): super().__init__() self.num_classes = num_classes self.dim = dim self.fc1 = nn.Linear(self.dim, 512) self...
ScaleLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ScaleLayer(nn.Module): def __init__(self, channels, skip_dims=2): super().__init__() self.scale = nn.Parameter(torch.ones(channels, *([1] * skip_dims))) def forward(self, net): return net * self.scale def extra_repr(self): return ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
JGU-VC/activation-pattern-analysis
ScaleLayer
false
589
[ "MIT" ]
0
14da42ad541ee4faf35d360a6e871fd44decd33d
https://github.com/JGU-VC/activation-pattern-analysis/tree/14da42ad541ee4faf35d360a6e871fd44decd33d
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, channels, skip_dims=2): super().__init__() self.scale = nn.Parameter(torch.ones(channels, *([1] * skip_dims))) def forward(self, net): return net * self.scale def extra_repr(self): return f'sha...
L1GradLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class L1GradLoss(nn.Module): def __init__(self, grad=False): super(L1GradLoss, self).__init__() self.grad = grad def forward(self, input, target): err = input - target loss = err.norm(p=1).div(err.numel()) if ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
JaguAroo/SRResCGAN
L1GradLoss
false
590
[ "MIT" ]
0
9aac612aff631f7fb9142e0a36de9559cfc1a62d
https://github.com/JaguAroo/SRResCGAN/tree/9aac612aff631f7fb9142e0a36de9559cfc1a62d
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): def __init__(self, grad=False): super().__init__() self.grad = grad def forward(self, input, target): err = input - target loss = err.norm(p=1).div(err.numel()) if self.grad: ...
Policy
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim import torch.optim import torch.autograd class Policy(nn.Module): def __init__(self, learning_rate, gamma, in_dim, out_dim): super(Policy, self).__init__() self.learning_rate = learning_rate sel...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
ChangQingAAS/Deep-Reinforcement-Learning
Policy
false
591
[ "MIT" ]
0
3bc1381c632b1730a48e63e972aea62086c4287c
https://github.com/ChangQingAAS/Deep-Reinforcement-Learning/tree/3bc1381c632b1730a48e63e972aea62086c4287c
import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim import torch.optim import torch.autograd class Model(nn.Module): def __init__(self, learning_rate, gamma, in_dim, out_dim): super().__init__() self.learning_rate = learning_rate self.gamma = gam...
AdaptiveConcatPool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import * from typing import Optional from torch import nn class AdaptiveConcatPool2d(nn.Module): """Layer that concats `AdaptiveAvgPool2d` and `AdaptiveMaxPool2d`.""" def __init__(self, sz: 'Optional[int]'=None): super(AdaptiveConcatPool2d, self).__init__() """Output ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from typing import * from typing import Optional from torch import nn assert_size_stride ...
JacobARose/image-utils
AdaptiveConcatPool2d
false
592
[ "MIT" ]
0
aa0e005c0b4df5198d188b074f4e21f8d8f97962
https://github.com/JacobARose/image-utils/tree/aa0e005c0b4df5198d188b074f4e21f8d8f97962
import torch from typing import * from typing import Optional from torch import nn class Model(nn.Module): """Layer that concats `AdaptiveAvgPool2d` and `AdaptiveMaxPool2d`.""" def __init__(self, sz: 'Optional[int]'=None): super().__init__() """Output will be 2*sz or 2 if sz is None""" ...
ILN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from torch.nn.parameter import Parameter import torch.utils.data class ILN(nn.Module): def __init__(self, num_features, eps=1e-05): super(ILN, self).__init__() self.eps = eps self.rho = Parameter(torch.Tensor(1, num_features, 1, 1)) self.gamma = ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn from torch.nn.parameter import Parameter import torch.uti...
JW9MsjwjnpdRLFw/RMT
ILN
false
593
[ "MIT" ]
0
a877fd78639a8d4c534d0373b9d0ad023e0fa2dd
https://github.com/JW9MsjwjnpdRLFw/RMT/tree/a877fd78639a8d4c534d0373b9d0ad023e0fa2dd
import torch import torch.nn as nn from torch.nn.parameter import Parameter import torch.utils.data class Model(nn.Module): def __init__(self, num_features, eps=1e-05): super().__init__() self.eps = eps self.rho = Parameter(torch.Tensor(1, num_features, 1, 1)) self.gamma = Paramet...
CircleLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import * from torch import nn import torch.nn.functional as F from torch import functional as F from torch.nn import functional as F class CircleLoss(nn.Module): """CircleLoss from `"Circle Loss: A Unified Perspective of Pair Similarity Optimization" <https://arxiv.org/pdf/2002.10857>`_...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
JacobARose/image-utils
CircleLoss
false
594
[ "MIT" ]
0
aa0e005c0b4df5198d188b074f4e21f8d8f97962
https://github.com/JacobARose/image-utils/tree/aa0e005c0b4df5198d188b074f4e21f8d8f97962
import torch from typing import * from torch import nn import torch.nn.functional as F from torch import functional as F from torch.nn import functional as F class Model(nn.Module): """CircleLoss from `"Circle Loss: A Unified Perspective of Pair Similarity Optimization" <https://arxiv.org/pdf/2002.10857>`_ pape...
BasicBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn def conv3x3(in_planes, out_planes, stride=1, dilation=1): return nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=dilation, dilation=dilation, bias=False) class BasicBlock(nn.Module): expansion = 1 def __init__(self, inplan...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn import t...
Jack12xl/scene-representation-networks
BasicBlock
false
595
[ "MIT" ]
0
2691b23c956cf188a1fe4c84a888b19871cac8f4
https://github.com/Jack12xl/scene-representation-networks/tree/2691b23c956cf188a1fe4c84a888b19871cac8f4
import torch from torch import nn import torch.nn def conv3x3(in_planes, out_planes, stride=1, dilation=1): return nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, padding=dilation, dilation=dilation, bias=False) class Model(nn.Module): expansion = 1 def __init__(self, inplanes, p...
ResnetBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils def actvn(x): out = nn.functional.leaky_relu(x, 0.2) return out class ResnetBlock(nn.Module): def __init__(self, fin, fout, fhidden=None, is_bias=True): super(ResnetBlock, self).__init__() ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn import torch.nn.parallel import torch.utils.data import tor...
IdanAzuri/MixMatch-pytorch
ResnetBlock
false
596
[ "MIT" ]
0
b8de2bc30c09e1256b92e0394403487fc4f90135
https://github.com/IdanAzuri/MixMatch-pytorch/tree/b8de2bc30c09e1256b92e0394403487fc4f90135
import torch from torch import nn import torch.nn.parallel import torch.utils.data import torch.utils def actvn(x): out = nn.functional.leaky_relu(x, 0.2) return out class Model(nn.Module): def __init__(self, fin, fout, fhidden=None, is_bias=True): super().__init__() self.is_bias = is_b...
GaussianFilter
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class GaussianFilter(nn.Module): def __init__(self, kernel_size=13, stride=1, padding=6): super(GaussianFilter, self).__init__() mean = (kernel_size - 1) / 2.0 variance = ((kernel_size - 1) / 6.0) ** 2.0 x_coord = torch.ar...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.utils.data assert_size_stride = torch._C._dyn...
JaguAroo/SRResCGAN
GaussianFilter
false
597
[ "MIT" ]
0
9aac612aff631f7fb9142e0a36de9559cfc1a62d
https://github.com/JaguAroo/SRResCGAN/tree/9aac612aff631f7fb9142e0a36de9559cfc1a62d
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): def __init__(self, kernel_size=13, stride=1, padding=6): super().__init__() mean = (kernel_size - 1) / 2.0 variance = ((kernel_size - 1) / 6.0) ** 2.0 x_coord = torch.arange(kernel_size) x_g...
RecurrentNeuralNetwork
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import Tensor from torch.functional import Tensor from torch import nn from typing import Tuple from typing import Any class RecurrentNeuralNetwork(nn.Module): def __init__(self, input_size, hidden_size, output_size): super().__init__() self.hidden_size = hidden_size ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
JThissen/machine_learning
RecurrentNeuralNetwork
false
598
[ "MIT" ]
0
82e2b003fb25111dc2d9ac1c1b2fd637e9f4fdbc
https://github.com/JThissen/machine_learning/tree/82e2b003fb25111dc2d9ac1c1b2fd637e9f4fdbc
import torch from torch import Tensor from torch.functional import Tensor from torch import nn from typing import Tuple from typing import Any class Model(nn.Module): def __init__(self, input_size, hidden_size, output_size): super().__init__() self.hidden_size = hidden_size self.input2out...
LayerNormConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn class LayerNormConv2d(nn.Module): def __init__(self, num_features, eps=1e-05, affine=True): super().__init__() self.num_features = num_features self.affine = affine self.eps = eps if self.affine: self.gamma = nn...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice from torch import nn import torch.nn assert_size_stride = torch._C._dynamo.guar...
Jack12xl/scene-representation-networks
LayerNormConv2d
false
599
[ "MIT" ]
0
2691b23c956cf188a1fe4c84a888b19871cac8f4
https://github.com/Jack12xl/scene-representation-networks/tree/2691b23c956cf188a1fe4c84a888b19871cac8f4
import torch from torch import nn import torch.nn class Model(nn.Module): def __init__(self, num_features, eps=1e-05, affine=True): super().__init__() self.num_features = num_features self.affine = affine self.eps = eps if self.affine: self.gamma = nn.Parameter...
TransformerLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from typing import Optional from torch import nn import torch.nn.functional as nnf class MlpTransformer(nn.Module): def __init__(self, in_dim, h_dim, out_d: 'Optional[int]'=None, act=nnf. relu, dropout=0.0): super().__init__() out_d = out_d if out_d is not None else in_dim ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
JAVI897/CLIP_prefix_caption
TransformerLayer
false
600
[ "MIT" ]
0
f4569891d01a5a790e9cdf850fb7feda3a0affc7
https://github.com/JAVI897/CLIP_prefix_caption/tree/f4569891d01a5a790e9cdf850fb7feda3a0affc7
import torch from typing import Optional from torch import nn import torch.nn.functional as nnf class MlpTransformer(nn.Module): def __init__(self, in_dim, h_dim, out_d: 'Optional[int]'=None, act=nnf. relu, dropout=0.0): super().__init__() out_d = out_d if out_d is not None else in_dim ...
MyLinear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn from torch.nn import functional as F import torchvision.transforms.functional as F import torch.nn.functional as F class MyLinear(nn.Module): """Linear layer with equalized learning rate and custom learning rate multiplier.""" def __init__(self, input_size, output_size, gain...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_st...
AnimeshKoratana/blurryface
MyLinear
false
601
[ "Apache-2.0" ]
0
c6cb5feec02f6d5af3acb1678336800390715d65
https://github.com/AnimeshKoratana/blurryface/tree/c6cb5feec02f6d5af3acb1678336800390715d65
import torch from torch import nn from torch.nn import functional as F import torchvision.transforms.functional as F import torch.nn.functional as F class Model(nn.Module): """Linear layer with equalized learning rate and custom learning rate multiplier.""" def __init__(self, input_size, output_size, gain=2 ...
KnowledgeDistillationLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from typing import * from torch import nn import torch.nn.functional as F from torch import functional as F from torch.nn import functional as F class KnowledgeDistillationLoss(nn.Module): def __init__(self, temperature=1): super().__init__() self.temperature = temperature def f...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math from typing import * f...
JacobARose/image-utils
KnowledgeDistillationLoss
false
602
[ "MIT" ]
0
aa0e005c0b4df5198d188b074f4e21f8d8f97962
https://github.com/JacobARose/image-utils/tree/aa0e005c0b4df5198d188b074f4e21f8d8f97962
import torch from typing import * from torch import nn import torch.nn.functional as F from torch import functional as F from torch.nn import functional as F class Model(nn.Module): def __init__(self, temperature=1): super().__init__() self.temperature = temperature def forward(self, student...
AddPositionEmbs
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from typing import * from torch import nn class AddPositionEmbs(nn.Module): """Adds (optionally learned) positional embeddings to the inputs.""" def __init__(self, num_patches: 'int', dim: 'int', dropout_rate: 'float'=0.0): super(AddPositionEmbs, self).__init__() self.pos...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from typing import * from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dyn...
JacobARose/image-utils
AddPositionEmbs
false
603
[ "MIT" ]
0
aa0e005c0b4df5198d188b074f4e21f8d8f97962
https://github.com/JacobARose/image-utils/tree/aa0e005c0b4df5198d188b074f4e21f8d8f97962
import torch from typing import * from torch import nn class Model(nn.Module): """Adds (optionally learned) positional embeddings to the inputs.""" def __init__(self, num_patches: 'int', dim: 'int', dropout_rate: 'float'=0.0): super().__init__() self.pos_embedding = nn.Parameter(torch...
SplAtConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.autograd import Function from torch.nn import Module import logging import torch import torch.utils.data import torch.distributed as dist from torch import nn import torch.nn.functional as F from torch.autograd.function import Function from torch.autograd import Function from torch.nn.modules.utils import _p...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
JHMeusener/detectron2-ResNeSt
SplAtConv2d
false
604
[ "Apache-2.0" ]
0
6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
https://github.com/JHMeusener/detectron2-ResNeSt/tree/6abab6fb9496a528f6aa2d4e1e27f3e7ceb42685
from torch.autograd import Function from torch.nn import Module import logging import torch import torch.utils.data import torch.distributed as dist from torch import nn import torch.nn.functional as F from torch.autograd.function import Function from torch.autograd import Function from torch.nn.modules.utils import _p...
cPReLU
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class cPReLU(nn.Module): def __init__(self, complex_axis=1): super(cPReLU, self).__init__() self.r_prelu = nn.PReLU() self.i_prelu = nn.PReLU() self.complex_axis = complex_axis def forward(self, inputs): real, imag = torch.chunk(inpu...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
JamesLiao714/FullSubNet
cPReLU
false
605
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, complex_axis=1): super().__init__() self.r_prelu = nn.PReLU() self.i_prelu = nn.PReLU() self.complex_axis = complex_axis def forward(self, inputs): real, imag = torch.chunk(inputs, 2, self.c...
VAE
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch.nn import functional as F from torch import nn import torch.nn class VAE(nn.Module): def __init__(self, in_ch, out_ch, hidden_ch=128): super(VAE, self).__init__() self.in_ch = in_ch self.out_ch = out_ch self.fc1 = nn.Linear(in_ch, hidden_ch) self.fc...
import torch from torch import device from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from...
Jack12xl/scene-representation-networks
VAE
false
606
[ "MIT" ]
0
2691b23c956cf188a1fe4c84a888b19871cac8f4
https://github.com/Jack12xl/scene-representation-networks/tree/2691b23c956cf188a1fe4c84a888b19871cac8f4
import torch from torch.nn import functional as F from torch import nn import torch.nn class Model(nn.Module): def __init__(self, in_ch, out_ch, hidden_ch=128): super().__init__() self.in_ch = in_ch self.out_ch = out_ch self.fc1 = nn.Linear(in_ch, hidden_ch) self.fc21 = nn...
RealConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class RealConv2d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), dilation=1, groups=1, causal=True, complex_axis=1): """ in_channels: real+imag ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
JamesLiao714/FullSubNet
RealConv2d
false
607
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), dilation=1, groups=1, causal=True, complex_axis=1): """ in_channels: real+imag ...
VAE
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class VAE(nn.Module): def __init__(self): super(VAE, self).__init__() self.fc1 = nn.Linear(784, 5000) self.fc2 = nn.Linear(5000, 5000) self.fc21 = nn.Linear(5000, 20) self.fc22 = nn....
import torch from torch import device from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from...
DanIulian/minigrid_rl
VAE
false
608
[ "MIT" ]
0
d7b59fd1d1e62fc99d5134c89f59c6ad16246cfa
https://github.com/DanIulian/minigrid_rl/tree/d7b59fd1d1e62fc99d5134c89f59c6ad16246cfa
import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.data class Model(nn.Module): def __init__(self): super().__init__() self.fc1 = nn.Linear(784, 5000) self.fc2 = nn.Linear(5000, 5000) self.fc21 = nn.Linear(5000, 20) self.fc22 = nn.Linear(...
rmse
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch class rmse(torch.nn.Module): def __init__(self): super(rmse, self).__init__() def forward(self, y_true, y_pred): mse = torch.mean((y_pred - y_true) ** 2, axis=-1) rmse = torch.sqrt(mse + 1e-07) return torch.mean(rmse) def get_inputs(): return [torch.rand([4...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_c...
JamesLiao714/FullSubNet
rmse
false
609
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch class Model(torch.nn.Module): def __init__(self): super().__init__() def forward(self, y_true, y_pred): mse = torch.mean((y_pred - y_true) ** 2, axis=-1) rmse = torch.sqrt(mse + 1e-07) return torch.mean(rmse) def get_inputs(): return [torch.rand([4, 4, 4, 4...
TV_L1Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class TV_L1Loss(nn.Module): def __init__(self, tv_loss_weight=1): super(TV_L1Loss, self).__init__() def forward(self, x): batch_size = x.size()[0] h_x = x.size()[2] w_x = x.size()[3] count_h = self.tensor_size...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn import torch.utils.data assert_size_stride = torch....
JaguAroo/SRResCGAN
TV_L1Loss
false
610
[ "MIT" ]
0
9aac612aff631f7fb9142e0a36de9559cfc1a62d
https://github.com/JaguAroo/SRResCGAN/tree/9aac612aff631f7fb9142e0a36de9559cfc1a62d
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): def __init__(self, tv_loss_weight=1): super().__init__() def forward(self, x): batch_size = x.size()[0] h_x = x.size()[2] w_x = x.size()[3] count_h = self.tensor_size(x[:, :, 1:, :]) ...
FocalLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class FocalLoss(nn.Module): def __init__(self, gamma=2): super().__init__() self.gamma = gamma def forward(self, logit, target, epoch=0): target = target.float() max_val = (-logit).clamp(min=0) loss = ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
Jason-George/irn
FocalLoss
false
611
[ "MIT" ]
0
b75441b5fb8080c1dbd8dbcb9b05720a4ceb2246
https://github.com/Jason-George/irn/tree/b75441b5fb8080c1dbd8dbcb9b05720a4ceb2246
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, gamma=2): super().__init__() self.gamma = gamma def forward(self, logit, target, epoch=0): target = target.float() max_val = (-logit).clamp(min=0) loss = logi...
TV_L1LOSS
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class TV_L1LOSS(nn.Module): def __init__(self): super(TV_L1LOSS, self).__init__() def forward(self, x, y): size = x.size() h_tv_diff = torch.abs(x[:, :, 1:, :] - x[:, :, :-1, :] - (y[:, :, 1 :, :] - y[:, :, :-1, :...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn import torch.utils.data assert_size_stride = torch....
JaguAroo/SRResCGAN
TV_L1LOSS
false
612
[ "MIT" ]
0
9aac612aff631f7fb9142e0a36de9559cfc1a62d
https://github.com/JaguAroo/SRResCGAN/tree/9aac612aff631f7fb9142e0a36de9559cfc1a62d
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): def __init__(self): super().__init__() def forward(self, x, y): size = x.size() h_tv_diff = torch.abs(x[:, :, 1:, :] - x[:, :, :-1, :] - (y[:, :, 1 :, :] - y[:, :, :-1, :])).sum() w...
RealConvTranspose2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class RealConvTranspose2d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), output_padding=(0, 0), groups=1): """ in_channels: real+imag out_channels: real+imag """ ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
JamesLiao714/FullSubNet
RealConvTranspose2d
false
613
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), output_padding=(0, 0), groups=1): """ in_channels: real+imag out_channels: real+imag """ super().__in...
DiscShiftLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class DiscShiftLoss(nn.Module): """Disc shift loss. Args: loss_weight (float, optional): Loss weight. Defaults to 1.0. """ def __init__(self, loss_weight=0.1): super().__init__() self.loss_weight = loss_weight def forward(self, ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Jason-Khan/mmediting
DiscShiftLoss
false
614
[ "Apache-2.0" ]
0
d187f95a675dff3eb975a575bd9278d643b5b645
https://github.com/Jason-Khan/mmediting/tree/d187f95a675dff3eb975a575bd9278d643b5b645
import torch import torch.nn as nn class Model(nn.Module): """Disc shift loss. Args: loss_weight (float, optional): Loss weight. Defaults to 1.0. """ def __init__(self, loss_weight=0.1): super().__init__() self.loss_weight = loss_weight def forward(self, x): ...
ComplexConvTranspose2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ComplexConvTranspose2d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), output_padding=(0, 0), causal=False, complex_axis=1, groups=1): """ in_channels: real+imag o...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
JamesLiao714/FullSubNet
ComplexConvTranspose2d
false
615
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), output_padding=(0, 0), causal=False, complex_axis=1, groups=1): """ in_channels: real+imag out_channels: real...
ComplexConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class ComplexConv2d(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), dilation=1, groups=1, causal=True, complex_axis=1): """ in_channels: real+imag ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
JamesLiao714/FullSubNet
ComplexConv2d
false
616
[ "MIT" ]
0
dad740bac35b5d7544c97740ae59101455acdc40
https://github.com/JamesLiao714/FullSubNet/tree/dad740bac35b5d7544c97740ae59101455acdc40
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size=(1, 1), stride=(1, 1), padding=(0, 0), dilation=1, groups=1, causal=True, complex_axis=1): """ in_channels: real+imag ...
CharbonnierLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import functools import torch from torch.nn import functional as F import torch.nn as nn def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Returns: Tensor: Reduced lo...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import functools from torch....
Jason-Khan/mmediting
CharbonnierLoss
false
617
[ "Apache-2.0" ]
0
d187f95a675dff3eb975a575bd9278d643b5b645
https://github.com/Jason-Khan/mmediting/tree/d187f95a675dff3eb975a575bd9278d643b5b645
import functools import torch from torch.nn import functional as F import torch.nn as nn def reduce_loss(loss, reduction): """Reduce loss as specified. Args: loss (Tensor): Elementwise loss tensor. reduction (str): Options are "none", "mean" and "sum". Returns: Tensor: Reduced lo...