entry_point
stringlengths
1
65
original_triton_code
stringlengths
4.5k
619k
python_code
stringlengths
208
60.9k
triton_code
stringlengths
1.15k
275k
repo_name
stringlengths
7
115
module_name
stringlengths
1
65
synthetic
bool
1 class
uuid
int64
0
18.5k
licenses
listlengths
1
6
stars
int64
0
19.8k
sha
stringlengths
40
40
repo_link
stringlengths
72
180
pytorch_code
stringlengths
200
4.05k
ConvBlockFixup
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class ConvBlockFixup(nn.Module): def __init__(self, filter_width, input_filters, nb_filters, dilation): super(ConvBlockFixup, self).__init__() self.filter_width = filter_width self.input_filters = input_filters self.nb_filters = nb_filters ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
OmarNajdi/dl-for-har
ConvBlockFixup
false
934
[ "MIT" ]
0
5e1b7c29caf2b41fcba106cd901c45d8f2d18429
https://github.com/OmarNajdi/dl-for-har/tree/5e1b7c29caf2b41fcba106cd901c45d8f2d18429
import torch from torch import nn class Model(nn.Module): def __init__(self, filter_width, input_filters, nb_filters, dilation): super().__init__() self.filter_width = filter_width self.input_filters = input_filters self.nb_filters = nb_filters self.dilation = dilation ...
MarginRankingLoss_learning_loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class MarginRankingLoss_learning_loss(nn.Module): """ Ranking loss as described in LPM paper inputs/targets are randomly permutated final target is a list of -1 and 1's -1 means the item in the i list is higher 1 means the item i...
import torch from torch import device import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guar...
Pepijnnn/MasterThesis
MarginRankingLoss_learning_loss
false
936
[ "MIT" ]
0
7ec831f5e55f5f181e0196fa78284e2846ce2e26
https://github.com/Pepijnnn/MasterThesis/tree/7ec831f5e55f5f181e0196fa78284e2846ce2e26
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): """ Ranking loss as described in LPM paper inputs/targets are randomly permutated final target is a list of -1 and 1's -1 means the item in the i list is higher 1 means the item in the j list is higher ...
LRN
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class LRN(nn.Module): def __init__(self, local_size=1, alpha=0.0001, beta=0.75, ACROSS_CHANNELS=False): super(LRN, self).__init__() self.ACROSS_CHANNELS = ACROSS_CHANNELS if self.ACROSS_CHANNELS: self.average = nn.AvgPool3d(kernel_siz...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
PengJingchao/DFNet
LRN
false
937
[ "MIT" ]
0
49e83501f81515aebca211351e315896da7afc54
https://github.com/PengJingchao/DFNet/tree/49e83501f81515aebca211351e315896da7afc54
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, local_size=1, alpha=0.0001, beta=0.75, ACROSS_CHANNELS=False): super().__init__() self.ACROSS_CHANNELS = ACROSS_CHANNELS if self.ACROSS_CHANNELS: self.average = nn.AvgPool3d(kernel_size=(loca...
ListNetLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class ListNetLoss(nn.Module): def __init__(self): super(ListNetLoss, self).__init__() return def forward(self, y_pred, y_true, eps=0.0005, padded_value_indicator=-1): """ ListNet loss introduced in "Learning t...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
Pepijnnn/MasterThesis
ListNetLoss
false
938
[ "MIT" ]
0
7ec831f5e55f5f181e0196fa78284e2846ce2e26
https://github.com/Pepijnnn/MasterThesis/tree/7ec831f5e55f5f181e0196fa78284e2846ce2e26
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() return def forward(self, y_pred, y_true, eps=0.0005, padded_value_indicator=-1): """ ListNet loss introduced in "Learning to Rank: From Pairwise A...
AsymmetricLossOptimized
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class AsymmetricLossOptimized(nn.Module): """ Notice - optimized version, minimizes memory allocation and gpu uploading, favors inplace operations""" def __init__(self, gamma_neg=4, gamma_pos=1, clip=0.05, eps=1e-08, disable_torch_grad_focal_loss=False): ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
Pepijnnn/MasterThesis
AsymmetricLossOptimized
false
939
[ "MIT" ]
0
7ec831f5e55f5f181e0196fa78284e2846ce2e26
https://github.com/Pepijnnn/MasterThesis/tree/7ec831f5e55f5f181e0196fa78284e2846ce2e26
import torch import torch.nn as nn class Model(nn.Module): """ Notice - optimized version, minimizes memory allocation and gpu uploading, favors inplace operations""" def __init__(self, gamma_neg=4, gamma_pos=1, clip=0.05, eps=1e-08, disable_torch_grad_focal_loss=False): super().__init__(...
HorizontalMaxPool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class HorizontalMaxPool2d(nn.Module): def __init__(self): super(HorizontalMaxPool2d, self).__init__() def forward(self, x): inp_size = x.size() return nn.functional.max_pool2d(input=x, kernel_size=(1, inp_size[3])) def get_inputs(): return [to...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Qidian213/NAIC2019
HorizontalMaxPool2d
false
940
[ "MIT" ]
0
23e05a8a096168ccfa4d1743467fdf78ffcaabba
https://github.com/Qidian213/NAIC2019/tree/23e05a8a096168ccfa4d1743467fdf78ffcaabba
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, x): inp_size = x.size() return nn.functional.max_pool2d(input=x, kernel_size=(1, inp_size[3])) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_...
BinaryLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class BinaryLoss(nn.Module): def __init__(self): super(BinaryLoss, self).__init__() def forward(self, pos_score, neg_score): pos_loss = -F.log_softmax(pos_score, dim=1)[:, 1] neg_loss = -F.log_softmax(neg_score, dim=1...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
PengJingchao/DFNet
BinaryLoss
false
941
[ "MIT" ]
0
49e83501f81515aebca211351e315896da7afc54
https://github.com/PengJingchao/DFNet/tree/49e83501f81515aebca211351e315896da7afc54
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() def forward(self, pos_score, neg_score): pos_loss = -F.log_softmax(pos_score, dim=1)[:, 1] neg_loss = -F.log_softmax(neg_score, dim=1)[:, 0] loss ...
HSigmoid
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.functional import torch.optim import torch.nn.parallel import torch.utils.data.distributed class HSigmoid(nn.Module): """ Applies the Hard-Sigmoid function element-wise. `"Searching for Mob...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.utils.data import torch import torch.nn as nn import torch.nn.functional imp...
PhelaPoscam/SRGAN-PyTorch
HSigmoid
false
942
[ "Apache-2.0" ]
0
c1c68707dbddd1130b2ea71023df748080bcbd52
https://github.com/PhelaPoscam/SRGAN-PyTorch/tree/c1c68707dbddd1130b2ea71023df748080bcbd52
import torch import torch.utils.data import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.functional import torch.optim import torch.nn.parallel import torch.utils.data.distributed class Model(nn.Module): """ Applies the Hard-Sigmoid function element-wise. `"Searching for Mobile...
Module_CharbonnierLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class Module_CharbonnierLoss(nn.Module): def __init__(self, epsilon=0.001): super(Module_CharbonnierLoss, self).__init__() self.epsilon = epsilon def forward(self, output, gt): return torch.mean(torch.sqrt((output - gt) ** 2 + self.epsilon ** 2)) ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert...
Pumpkin123709/LBEC
Module_CharbonnierLoss
false
944
[ "MIT" ]
0
18661faa35769f731847e0226ff601754e134668
https://github.com/Pumpkin123709/LBEC/tree/18661faa35769f731847e0226ff601754e134668
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, epsilon=0.001): super().__init__() self.epsilon = epsilon def forward(self, output, gt): return torch.mean(torch.sqrt((output - gt) ** 2 + self.epsilon ** 2)) def get_inputs(): return [torch.rand([4, ...
Attention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Attention(nn.Module): """ Applies attention mechanism on the `context` using the `query`. **Thank you** to IBM for their initial implementation of :class:`Attention`. Here is their `License <https://github.com/IBM/pytorch-seq2seq/blob/master/LICENSE>`__. ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
PattynR/PyTorch-NLP
Attention
false
945
[ "BSD-3-Clause" ]
0
8995774abf3734db6da174425843d883face5218
https://github.com/PattynR/PyTorch-NLP/tree/8995774abf3734db6da174425843d883face5218
import torch import torch.nn as nn class Model(nn.Module): """ Applies attention mechanism on the `context` using the `query`. **Thank you** to IBM for their initial implementation of :class:`Attention`. Here is their `License <https://github.com/IBM/pytorch-seq2seq/blob/master/LICENSE>`__. Args...
Reg_layer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class Reg_layer(nn.Module): """ modified by Zylo117 """ def __init__(self, in_channels, out_channels): super().__init__() self.conv = nn.Conv2d(in_channels, in_channels, kernel_size=3, padding=1, bias=True) self.header = nn.Conv2d(...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
Peiiii/detro
Reg_layer
false
946
[ "MIT" ]
0
26d74468d7554dc20b2a2daf7ec5009302c820f2
https://github.com/Peiiii/detro/tree/26d74468d7554dc20b2a2daf7ec5009302c820f2
import torch from torch import nn class Model(nn.Module): """ modified by Zylo117 """ def __init__(self, in_channels, out_channels): super().__init__() self.conv = nn.Conv2d(in_channels, in_channels, kernel_size=3, padding=1, bias=True) self.header = nn.Conv2d(in_c...
Classifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Classifier(nn.Module): def __init__(self, num_inputs1, num_inputs2): super().__init__() self.network = nn.Bilinear(num_inputs1, num_inputs2, 1) def forward(self, x1, x2): return self.network(x1, x2) def get_inputs(): return [torch.rand([...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride reinterpret_tensor = torch._C._dynamo.guards._reinterp...
Project-Agni/Detection
Classifier
false
947
[ "MIT" ]
0
6b2c8ec25f8bd2bd15995d67f2808352cec9e2af
https://github.com/Project-Agni/Detection/tree/6b2c8ec25f8bd2bd15995d67f2808352cec9e2af
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, num_inputs1, num_inputs2): super().__init__() self.network = nn.Bilinear(num_inputs1, num_inputs2, 1) def forward(self, x1, x2): return self.network(x1, x2) def get_inputs(): return [torch.rand([4, 4,...
DQNetwork
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F import torch.nn as nn class Prediction(nn.Module): """Defines the prediction module as an ANN""" def __init__(self, state_size: 'int', action_size: 'int', fc1: 'int'=24, fc2: 'int'=24): super(Prediction, self).__init__() self.state_size = s...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn.functional as...
QasimWani/EARL
DQNetwork
false
948
[ "MIT" ]
0
05c303335e67903380771c4954a5317bd46fc0e7
https://github.com/QasimWani/EARL/tree/05c303335e67903380771c4954a5317bd46fc0e7
import torch import torch.nn.functional as F import torch.nn as nn class Prediction(nn.Module): """Defines the prediction module as an ANN""" def __init__(self, state_size: 'int', action_size: 'int', fc1: 'int'=24, fc2: 'int'=24): super().__init__() self.state_size = state_size ...
hsigmoid
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class hsigmoid(nn.Module): def forward(self, x): out = F.relu6(x + 3, inplace=True) / 6 return out def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], {}]
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Qidian213/NAIC2019
hsigmoid
false
949
[ "MIT" ]
0
23e05a8a096168ccfa4d1743467fdf78ffcaabba
https://github.com/Qidian213/NAIC2019/tree/23e05a8a096168ccfa4d1743467fdf78ffcaabba
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def forward(self, x): out = F.relu6(x + 3, inplace=True) / 6 return out def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []
SeparableConv2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class SeparableConv2d(torch.nn.Module): def __init__(self, in_channels, out_channels, kernel_size=1, stride=1, padding=0, dilation=1, bias=False): super(SeparableConv2d, self).__init__() self.conv1 = torch.nn.Conv2d(in_channels, in_channels, kernel_size, stride, p...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream assert_size_stride = torch._C._dynamo.guards.assert_size_stride @triton.jit de...
Pumpkin123709/LBEC
SeparableConv2d
false
950
[ "MIT" ]
0
18661faa35769f731847e0226ff601754e134668
https://github.com/Pumpkin123709/LBEC/tree/18661faa35769f731847e0226ff601754e134668
import torch class Model(torch.nn.Module): def __init__(self, in_channels, out_channels, kernel_size=1, stride=1, padding=0, dilation=1, bias=False): super().__init__() self.conv1 = torch.nn.Conv2d(in_channels, in_channels, kernel_size, stride, padding, dilation, groups=in_cha...
InputProjectionA
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data class InputProjectionA(nn.Module): """ This class projects the input image to the same spatial dimensions as the feature map. For example, if the input image is 512 x512 x3 and spatial dimensions of feature map size are 56...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.nn.parallel import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty...
PhillipHuang2017/ext_portrait_segmentation
InputProjectionA
false
951
[ "MIT" ]
0
6d0cec0a953dacbc94a01ea8b719feb687b7c029
https://github.com/PhillipHuang2017/ext_portrait_segmentation/tree/6d0cec0a953dacbc94a01ea8b719feb687b7c029
import torch import torch.nn as nn import torch.nn.parallel import torch.utils.data class Model(nn.Module): """ This class projects the input image to the same spatial dimensions as the feature map. For example, if the input image is 512 x512 x3 and spatial dimensions of feature map size are 56x56xF, then...
SelfAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class SelfAttention(nn.Module): """ Self attention Layer""" def __init__(self, in_dim, activation): super(SelfAttention, self).__init__() self.chanel_in = in_dim self.activation = activation self.query_conv = nn.Conv2d(in_channels=in_dim, out_...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
PauPerezT/EmoSSpeech
SelfAttention
false
952
[ "Apache-2.0" ]
0
168a951a838d0bfb838e7d0e3f6895bff68164a4
https://github.com/PauPerezT/EmoSSpeech/tree/168a951a838d0bfb838e7d0e3f6895bff68164a4
import torch from torch import nn class Model(nn.Module): """ Self attention Layer""" def __init__(self, in_dim, activation): super().__init__() self.chanel_in = in_dim self.activation = activation self.query_conv = nn.Conv2d(in_channels=in_dim, out_channels=in_dim // ...
Environment
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F import torch.nn as nn class Environment(nn.Module): """Defines the Environment module as an ANN""" def __init__(self, state_size: 'int', action_size: 'int', fc1: 'int'=24, fc2: 'int'=24): super(Environment, self).__init__() self.state_size ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
QasimWani/EARL
Environment
false
953
[ "MIT" ]
0
05c303335e67903380771c4954a5317bd46fc0e7
https://github.com/QasimWani/EARL/tree/05c303335e67903380771c4954a5317bd46fc0e7
import torch import torch.nn.functional as F import torch.nn as nn class Model(nn.Module): """Defines the Environment module as an ANN""" def __init__(self, state_size: 'int', action_size: 'int', fc1: 'int'=24, fc2: 'int'=24): super().__init__() self.state_size = state_size se...
attention2d
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class attention2d(nn.Module): def __init__(self, in_planes, ratios, K, temperature, init_weight=True): super(attention2d, self).__init__() assert temperature % 3 == 1 self.avgpool = nn.AdaptiveAvgPool2d(1) if in_pl...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
PengJingchao/DFNet
attention2d
false
954
[ "MIT" ]
0
49e83501f81515aebca211351e315896da7afc54
https://github.com/PengJingchao/DFNet/tree/49e83501f81515aebca211351e315896da7afc54
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, in_planes, ratios, K, temperature, init_weight=True): super().__init__() assert temperature % 3 == 1 self.avgpool = nn.AdaptiveAvgPool2d(1) if in_planes != 3: ...
ShearY
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class ShearY(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M - 180 def forward(self, img): return ttf.functional.affine(img, 0, [0, 0], 1, [0, self.ang...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Hayoung93/UDA
ShearY
false
955
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M - 180 def forward(self, img): return ttf.functional.affine(img, 0, [0, 0], 1, [0, self.angl...
fully_conv_layer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class fully_conv_layer(nn.Module): def __init__(self, c): super(fully_conv_layer, self).__init__() self.conv = nn.Conv2d(c, 1, 1) def forward(self, x): return self.conv(x) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inpu...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
QiweiMa-LL/STAGCN
fully_conv_layer
false
956
[ "MIT" ]
0
c6889c845ac7fcba4419b2727022a599981f2a54
https://github.com/QiweiMa-LL/STAGCN/tree/c6889c845ac7fcba4419b2727022a599981f2a54
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, c): super().__init__() self.conv = nn.Conv2d(c, 1, 1) def forward(self, x): return self.conv(x) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [4]
HSwish
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.functional import torch.optim import torch.nn.parallel import torch.utils.data.distributed class HSwish(nn.Module): """ Applies the Hard-Swish function element-wise. `"Searching for MobileN...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.utils.data import torch import torch.nn as nn import torch.nn.functional imp...
PhelaPoscam/SRGAN-PyTorch
HSwish
false
957
[ "Apache-2.0" ]
0
c1c68707dbddd1130b2ea71023df748080bcbd52
https://github.com/PhelaPoscam/SRGAN-PyTorch/tree/c1c68707dbddd1130b2ea71023df748080bcbd52
import torch import torch.utils.data import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.functional import torch.optim import torch.nn.parallel import torch.utils.data.distributed class Model(nn.Module): """ Applies the Hard-Swish function element-wise. `"Searching for MobileNe...
TranslateX
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class TranslateX(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): try: max_size = img.size()[0] except TypeError: max_size = img.size()...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Hayoung93/UDA
TranslateX
false
958
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): try: max_size = img.size()[0] except TypeError: max_size = img.size()[0] ...
Rotate
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class Rotate(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M def forward(self, img): return ttf.functional.rotate(img, self.angle) def get_inputs(): ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Hayoung93/UDA
Rotate
false
959
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M def forward(self, img): return ttf.functional.rotate(img, self.angle) def get_inputs(): ...
hswish
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class hswish(nn.Module): def forward(self, x): out = x * F.relu6(x + 3, inplace=True) / 6 return out def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return [[], {}]
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Qidian213/NAIC2019
hswish
false
960
[ "MIT" ]
0
23e05a8a096168ccfa4d1743467fdf78ffcaabba
https://github.com/Qidian213/NAIC2019/tree/23e05a8a096168ccfa4d1743467fdf78ffcaabba
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def forward(self, x): out = x * F.relu6(x + 3, inplace=True) / 6 return out def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inputs(): return []
TranslateY
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class TranslateY(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): try: max_size = img.size()[1] except TypeError: max_size = img.size()...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Hayoung93/UDA
TranslateY
false
961
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): try: max_size = img.size()[1] except TypeError: max_size = img.size()[1] ...
ListMLELoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class ListMLELoss(nn.Module): def __init__(self): super(ListMLELoss, self).__init__() return def forward(self, y_pred, y_true, eps=1e-05, padded_value_indicator=-1): """ ListMLE loss introduced in "Listwise Approach to Learning to Rank - The...
import torch from torch import device import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_ma...
Pepijnnn/MasterThesis
ListMLELoss
false
962
[ "MIT" ]
0
7ec831f5e55f5f181e0196fa78284e2846ce2e26
https://github.com/Pepijnnn/MasterThesis/tree/7ec831f5e55f5f181e0196fa78284e2846ce2e26
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() return def forward(self, y_pred, y_true, eps=1e-05, padded_value_indicator=-1): """ ListMLE loss introduced in "Listwise Approach to Learning to Rank - Theory and Algorithm". ...
DiceLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class DiceLoss(nn.Module): def __init__(self): super(DiceLoss, self).__init__() def forward(self, input, target): N = target.size(0) smooth = 1 input_flat = input.view(N, -1) target_flat = target.view(N, -1) intersection = in...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Qsingle/MedicalImage
DiceLoss
false
963
[ "MIT" ]
0
a5020d7d2266669a4d6ffec224430e8b25cc1dfc
https://github.com/Qsingle/MedicalImage/tree/a5020d7d2266669a4d6ffec224430e8b25cc1dfc
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, input, target): N = target.size(0) smooth = 1 input_flat = input.view(N, -1) target_flat = target.view(N, -1) intersection = input_flat * target...
DummyLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class DummyLayer(torch.nn.Module): def __init__(self): super().__init__() self.dummy = torch.nn.Parameter(torch.ones(1, dtype=torch.float32)) def forward(self, x): return x + self.dummy - self.dummy def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda @triton.j...
RICE-EIC/Early-Bird-GCN
DummyLayer
false
964
[ "Apache-2.0" ]
0
25a80b23f2ecfc46ffe00b1cf0e06052b32aad0f
https://github.com/RICE-EIC/Early-Bird-GCN/tree/25a80b23f2ecfc46ffe00b1cf0e06052b32aad0f
import torch class Model(torch.nn.Module): def __init__(self): super().__init__() self.dummy = torch.nn.Parameter(torch.ones(1, dtype=torch.float32)) def forward(self, x): return x + self.dummy - self.dummy def get_inputs(): return [torch.rand([4, 4, 4, 4])] def get_init_inpu...
DiceLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class DiceLoss(nn.Module): """ Criterion that computes Sørensen-Dice Coefficient loss. https://en.wikipedia.org/wiki/S%C3%B8rensen%E2%80%93Dice_coefficient """ def __init__(self): super().__init__() self.smooth = 1.0 def forward(self, input...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Quentin18/road-segmentation
DiceLoss
false
965
[ "MIT" ]
0
9d212c80fa3f6926c431847337d2ca38ec96b614
https://github.com/Quentin18/road-segmentation/tree/9d212c80fa3f6926c431847337d2ca38ec96b614
import torch import torch.nn as nn class Model(nn.Module): """ Criterion that computes Sørensen-Dice Coefficient loss. https://en.wikipedia.org/wiki/S%C3%B8rensen%E2%80%93Dice_coefficient """ def __init__(self): super().__init__() self.smooth = 1.0 def forward(self, input, t...
ShearX
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class ShearX(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M - 180 def forward(self, img): return ttf.functional.affine(img, 0, [0, 0], 1, [self.angle,...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as ...
Hayoung93/UDA
ShearX
false
966
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M self.angle = 359 / 10 * self.M - 180 def forward(self, img): return ttf.functional.affine(img, 0, [0, 0], 1, [self.angle, ...
Convolution
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as fn from torch.nn.parameter import Parameter import torch.nn def to_pair(data): """Converts a single or a tuple of data into a pair. If the data is a tuple with more than two elements, it selects the first two of them. In case of single data, it dup...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn from torch.nn.parameter import Parameter import torch.nn a...
R1704/SpeechRecognitionSNN
Convolution
false
967
[ "MIT" ]
0
4b788d1bd20d8ce201da6da8b200b3ca722c7efa
https://github.com/R1704/SpeechRecognitionSNN/tree/4b788d1bd20d8ce201da6da8b200b3ca722c7efa
import torch import torch.nn as nn import torch.nn.functional as fn from torch.nn.parameter import Parameter import torch.nn def to_pair(data): """Converts a single or a tuple of data into a pair. If the data is a tuple with more than two elements, it selects the first two of them. In case of single data, it dup...
HighwayLayer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators class HighwayLayer(nn.Module): def __init__(self, input_dim, transform_activation=F.relu, gate_activation=F.softmax, gate_bias=-2): super().__init__() ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
ROCmSoftwarePlatform/translate
HighwayLayer
false
968
[ "BSD-3-Clause" ]
0
32a6380d914ebe1a6c38c4992aac9600ed3d9810
https://github.com/ROCmSoftwarePlatform/translate/tree/32a6380d914ebe1a6c38c4992aac9600ed3d9810
import torch import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators class Model(nn.Module): def __init__(self, input_dim, transform_activation=F.relu, gate_activation=F.softmax, gate_bias=-2): super().__init__() self.h...
Attention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.utils.data import torch.utils.checkpoint class Attention(nn.Module): def __init__(self, in_size, hidden_size): super(Attention, self).__init__() self.hidden = nn.Linear(in_size, hidden_size) nn.init.orthogonal_(self.hidden.weight.data) ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
MarvinLvn/platalea
Attention
false
969
[ "Apache-2.0" ]
0
31def0813c90a3259f86f7d86cb576cd66dca3fe
https://github.com/MarvinLvn/platalea/tree/31def0813c90a3259f86f7d86cb576cd66dca3fe
import torch import torch.nn as nn import torch.utils.data import torch.utils.checkpoint class Model(nn.Module): def __init__(self, in_size, hidden_size): super().__init__() self.hidden = nn.Linear(in_size, hidden_size) nn.init.orthogonal_(self.hidden.weight.data) self.out = nn.Li...
Generator
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Generator(nn.Module): """Define standard linear + softmax generation step.""" def __init__(self, size, vocab): super(Generator, self).__init__() self.size = size self.proj = nn.Linear(self.size, vocab) def forward(self, x): sliced_...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_s...
QuLog1/QuLog
Generator
false
970
[ "Apache-2.0" ]
0
121f3a8c6f5ee60cde771c36b9eef823a1b2597a
https://github.com/QuLog1/QuLog/tree/121f3a8c6f5ee60cde771c36b9eef823a1b2597a
import torch import torch.nn as nn class Model(nn.Module): """Define standard linear + softmax generation step.""" def __init__(self, size, vocab): super().__init__() self.size = size self.proj = nn.Linear(self.size, vocab) def forward(self, x): sliced_x = x[:, 0, :] ...
CosLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.utils.data class CosLoss(nn.Module): def __init__(self, factor=6e-07, havesum=True, havemax=True): super(CosLoss, self).__init__() self.factor = factor self.havesum = havesum self.havemax = havemax def forward(self, w): m...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn import torch.utils.data assert_size_stride = torch._C._dynamo.guards...
PatrickGui/Face_Pytorch
CosLoss
false
971
[ "Apache-2.0" ]
0
ff5b820ca3978883f7cf95f0209fba3ee958c939
https://github.com/PatrickGui/Face_Pytorch/tree/ff5b820ca3978883f7cf95f0209fba3ee958c939
import torch from torch import nn import torch.utils.data class Model(nn.Module): def __init__(self, factor=6e-07, havesum=True, havemax=True): super().__init__() self.factor = factor self.havesum = havesum self.havemax = havemax def forward(self, w): mask = torch.one...
CumulativeLinkLoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import numpy as np from torch import nn from typing import Optional import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed def _reduction(loss: 'torch.Tensor', reduction: 'str') ->torch.Tensor: """ Reduce loss Parameters ---------- loss...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import numpy as np fro...
Ramstein/Retinopathy2
CumulativeLinkLoss
false
972
[ "MIT" ]
0
669e74206c466e6351d4e3df6087c6aa39b5c6c2
https://github.com/Ramstein/Retinopathy2/tree/669e74206c466e6351d4e3df6087c6aa39b5c6c2
import torch import numpy as np from torch import nn from typing import Optional import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed def _reduction(loss: 'torch.Tensor', reduction: 'str') ->torch.Tensor: """ Reduce loss Parameters ---------- loss...
CustomBatchNormAutograd
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class CustomBatchNormAutograd(nn.Module): """ This nn.module implements a custom version of the batch norm operation for MLPs. The operations called in self.forward track the history if the input tensors have the flag requires_grad set to True. """ def __ini...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
RaymondKoopmanschap/DL_assignment_code
CustomBatchNormAutograd
false
973
[ "MIT" ]
0
68b3290be9fbd6c55433a7585e2cfa18e0f35f5c
https://github.com/RaymondKoopmanschap/DL_assignment_code/tree/68b3290be9fbd6c55433a7585e2cfa18e0f35f5c
import torch import torch.nn as nn class Model(nn.Module): """ This nn.module implements a custom version of the batch norm operation for MLPs. The operations called in self.forward track the history if the input tensors have the flag requires_grad set to True. """ def __init__(self, n_neuron...
MaxPool
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class MaxPool(nn.Module): def __init__(self, kernel_size, stride=1, padding=1, zero_pad=False): super(MaxPool, self).__init__() self.zero_pad = nn.ZeroPad2d((1...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data ...
Ramstein/Retinopathy2
MaxPool
false
974
[ "MIT" ]
0
669e74206c466e6351d4e3df6087c6aa39b5c6c2
https://github.com/Ramstein/Retinopathy2/tree/669e74206c466e6351d4e3df6087c6aa39b5c6c2
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): def __init__(self, kernel_size, stride=1, padding=1, zero_pad=False): super().__init__() self.zero_pad = nn.ZeroPad2d((1, 0, 1, 0)) if ...
LinearEnsemble
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch as T import torch.nn as nn class LinearEnsemble(nn.Module): __constants__ = ['in_features', 'out_features'] ensemble_size: 'int' in_features: 'int' out_features: 'int' weight: 'T.Tensor' def __init__(self, ensemble_size: 'int', in_features: 'int', out_feature...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch as T import torch.nn as nn assert_size_stride = torch._C._dynamo.gu...
RamiSketcher/AMMI-RL
LinearEnsemble
false
975
[ "MIT" ]
0
6d51587ff4d5dc14cba87fca561bd7b340b44586
https://github.com/RamiSketcher/AMMI-RL/tree/6d51587ff4d5dc14cba87fca561bd7b340b44586
import torch import torch as T import torch.nn as nn class Model(nn.Module): __constants__ = ['in_features', 'out_features'] ensemble_size: 'int' in_features: 'int' out_features: 'int' weight: 'T.Tensor' def __init__(self, ensemble_size: 'int', in_features: 'int', out_features: 'int',...
WordPredictor
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators class WordPredictor(nn.Module): def __init__(self, encoder_output_dim, hidden_dim, output_dim, topk_labels_per_source_token=None, use_self_attention=False): s...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn.functional as...
ROCmSoftwarePlatform/translate
WordPredictor
false
976
[ "BSD-3-Clause" ]
0
32a6380d914ebe1a6c38c4992aac9600ed3d9810
https://github.com/ROCmSoftwarePlatform/translate/tree/32a6380d914ebe1a6c38c4992aac9600ed3d9810
import torch import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators class Model(nn.Module): def __init__(self, encoder_output_dim, hidden_dim, output_dim, topk_labels_per_source_token=None, use_self_attention=False): super()._...
GlobalAvgPool2d
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data class GlobalAvgPool2d(nn.Module): def __init__(self): """Global average pooling over the input's spatial dimensions""" super(GlobalAvgPool2d, self).__init__() def forward(self, inputs): in_size = inputs.size() return ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.utils.data assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C....
Rming/Self-Correction-Human-Parsing
GlobalAvgPool2d
false
977
[ "MIT" ]
0
c2b711c0a11f3980a8bf4c7a2acf85d80732620a
https://github.com/Rming/Self-Correction-Human-Parsing/tree/c2b711c0a11f3980a8bf4c7a2acf85d80732620a
import torch import torch.nn as nn import torch.utils.data class Model(nn.Module): def __init__(self): """Global average pooling over the input's spatial dimensions""" super().__init__() def forward(self, inputs): in_size = inputs.size() return inputs.view((in_size[0], in_siz...
CustomBatchNormManualModule
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class CustomBatchNormManualFunction(torch.autograd.Function): """ This torch.autograd.Function implements a functional custom version of the batch norm operation for MLPs. Using torch.autograd.Function allows you to write a custom backward function. The function will...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
RaymondKoopmanschap/DL_assignment_code
CustomBatchNormManualModule
false
978
[ "MIT" ]
0
68b3290be9fbd6c55433a7585e2cfa18e0f35f5c
https://github.com/RaymondKoopmanschap/DL_assignment_code/tree/68b3290be9fbd6c55433a7585e2cfa18e0f35f5c
import torch import torch.nn as nn class CustomBatchNormManualFunction(torch.autograd.Function): """ This torch.autograd.Function implements a functional custom version of the batch norm operation for MLPs. Using torch.autograd.Function allows you to write a custom backward function. The function will...
Encoder
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Encoder(nn.Module): def __init__(self, input_dim, hidden_dim, latent_dim): super(Encoder, self).__init__() self.FC_input = nn.Linear(input_dim, hidden_dim) self.FC_mean = nn.Linear(hidden_dim, latent_dim) self.FC_var = nn.Linear(hidden_dim,...
import torch from torch import device from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from...
RasmusJuul/dtu_mlops
Encoder
false
979
[ "Apache-2.0" ]
0
98bca082067aa7575bb8e8193991723d474f0850
https://github.com/RasmusJuul/dtu_mlops/tree/98bca082067aa7575bb8e8193991723d474f0850
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, input_dim, hidden_dim, latent_dim): super().__init__() self.FC_input = nn.Linear(input_dim, hidden_dim) self.FC_mean = nn.Linear(hidden_dim, latent_dim) self.FC_var = nn.Linear(hidden_dim, latent_dim) ...
Brightness
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torchvision import transforms as ttf class Brightness(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): return ttf.functional.adjust_brightness(img, self.M / 5.0) def get_inputs(): return [torch.rand...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Hayoung93/UDA
Brightness
false
980
[ "Apache-2.0" ]
0
a587b01c76141d64e7cead55b62e0f3ed75890bf
https://github.com/Hayoung93/UDA/tree/a587b01c76141d64e7cead55b62e0f3ed75890bf
import torch import torch.nn as nn from torchvision import transforms as ttf class Model(nn.Module): def __init__(self, M): super().__init__() self.M = M def forward(self, img): return ttf.functional.adjust_brightness(img, self.M / 5.0) def get_inputs(): return [torch.rand([4, ...
RegKappa
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch from torch.nn.modules.loss import _Loss import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class RegKappa(_Loss): def __init__(self, ignore_index=None): super(RegKappa, self).__init__() self.min = min self.max = max ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice from torch.nn.modules.loss i...
Ramstein/Retinopathy2
RegKappa
false
981
[ "MIT" ]
0
669e74206c466e6351d4e3df6087c6aa39b5c6c2
https://github.com/Ramstein/Retinopathy2/tree/669e74206c466e6351d4e3df6087c6aa39b5c6c2
import torch from torch.nn.modules.loss import _Loss import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(_Loss): def __init__(self, ignore_index=None): super().__init__() self.min = min self.max = max self.ignore_inde...
ScaledL2Norm
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.onnx import torch import torch.nn as nn import torch.nn.functional as F class ScaledL2Norm(nn.Module): def __init__(self, in_channels, initial_scale): super(ScaledL2Norm, self).__init__() self.in_channels = in_channels self.scale = nn.Parameter(torch.Tensor(in_ch...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice import torch.onnx import tor...
Richard-cpu2333/tx2dl
ScaledL2Norm
false
982
[ "Apache-2.0" ]
0
985d9f9f24004271e85745a49252ab9922aec655
https://github.com/Richard-cpu2333/tx2dl/tree/985d9f9f24004271e85745a49252ab9922aec655
import torch import torch.onnx import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, in_channels, initial_scale): super().__init__() self.in_channels = in_channels self.scale = nn.Parameter(torch.Tensor(in_channels)) self.ini...
CosineLinear
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter class CosineLinear(nn.Module): def __init__(self, in_features, out_features): super(CosineLinear, self).__init__() self.in_features = in_features self.out_features = out_features self...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
QingquanBao/Spear-Shield
CosineLinear
false
983
[ "Apache-2.0" ]
0
d57b8f4412c3d651b6f7e056c9c45cfd0dc950c3
https://github.com/QingquanBao/Spear-Shield/tree/d57b8f4412c3d651b6f7e056c9c45cfd0dc950c3
import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter class Model(nn.Module): def __init__(self, in_features, out_features): super().__init__() self.in_features = in_features self.out_features = out_features self.weight = Parameter(torch...
MultiheadAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import math import torch import numpy as np import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators def combine_heads(X): """ Combine heads (the inverse of split heads): 1) Transpose X from (batch size, nheads, sequence length, d_head) ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
ROCmSoftwarePlatform/translate
MultiheadAttention
false
984
[ "BSD-3-Clause" ]
0
32a6380d914ebe1a6c38c4992aac9600ed3d9810
https://github.com/ROCmSoftwarePlatform/translate/tree/32a6380d914ebe1a6c38c4992aac9600ed3d9810
import math import torch import numpy as np import torch.nn.functional as F import torch.nn as nn import torch.jit import torch.jit.quantized import torch.onnx.operators def combine_heads(X): """ Combine heads (the inverse of split heads): 1) Transpose X from (batch size, nheads, sequence length, d_head) ...
LogisticCumulativeLink
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class LogisticCumulativeLink(nn.Module): """ Converts a single number to the proportional odds of belonging to a class. Parameters ---------- num_classes : int ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed assert_size_str...
Ramstein/Retinopathy2
LogisticCumulativeLink
false
985
[ "MIT" ]
0
669e74206c466e6351d4e3df6087c6aa39b5c6c2
https://github.com/Ramstein/Retinopathy2/tree/669e74206c466e6351d4e3df6087c6aa39b5c6c2
import torch from torch import nn import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed class Model(nn.Module): """ Converts a single number to the proportional odds of belonging to a class. Parameters ---------- num_classes : int Number of...
Beta
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.nn.functional as F class BoundedBeta(torch.distributions.Beta): def log_prob(self, x): return super().log_prob((x + 1) / 2) class Beta(nn.Module): def __init__(self, action_dim): super(Beta, self).__init__() self.action_dim = action_d...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torch.nn as nn assert_size_stride = torch._C._dynamo.gu...
RohanPankaj/apex
Beta
false
986
[ "MIT" ]
0
74e96386bf9446d1179106d6d65ea0368c1b5b27
https://github.com/RohanPankaj/apex/tree/74e96386bf9446d1179106d6d65ea0368c1b5b27
import torch import torch.nn as nn import torch.nn.functional as F class BoundedBeta(torch.distributions.Beta): def log_prob(self, x): return super().log_prob((x + 1) / 2) class Model(nn.Module): def __init__(self, action_dim): super().__init__() self.action_dim = action_dim d...
L2Norm
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class L2Norm(nn.Module): def __init__(self, n_channels, scale=1.0): super(L2Norm, self).__init__() self.n_channels = n_channels self.scale = scale self.eps = 1e-10 self.weight = nn.Parameter(torch.Tensor(self.n_channels)) self.wei...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_...
Rocketbase-AI/rockets-s3fd
L2Norm
false
987
[ "MIT" ]
0
40d978270a6b3ba2d397217ede0c735712814250
https://github.com/Rocketbase-AI/rockets-s3fd/tree/40d978270a6b3ba2d397217ede0c735712814250
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, n_channels, scale=1.0): super().__init__() self.n_channels = n_channels self.scale = scale self.eps = 1e-10 self.weight = nn.Parameter(torch.Tensor(self.n_channels)) self.weight.data *= 0...
RMSLELoss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class RMSLELoss(nn.Module): def __init__(self): super().__init__() self.mse = nn.MSELoss() def forward(self, pred, actual): return torch.sqrt(self.mse(torch.log(pred + 1), torch.log(actual + 1))) def get_inputs(): return [torch.rand([4, 4, 4, ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math import torc...
RosarioAndolina/psychXRF
RMSLELoss
false
988
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() self.mse = nn.MSELoss() def forward(self, pred, actual): return torch.sqrt(self.mse(torch.log(pred + 1), torch.log(actual + 1))) def get_inputs(): return [torch.rand([4, 4, 4, 4]),...
SVIGlobalMeanPool2D
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class SVIGlobalMeanPool2D(nn.Module): """ Expects :param x: [examples, samples, channels, H, W] :return: [examples, samples, channels] """ def __init__(self): super(SVIGlobalMeanPool2D, self).__init__() def forward(self, x): x = x.mean(4...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
RomanShen/radial-bnn
SVIGlobalMeanPool2D
false
989
[ "Apache-2.0" ]
0
7c8bc85397c1461a6fd5ea9adf0631f9ade27f6c
https://github.com/RomanShen/radial-bnn/tree/7c8bc85397c1461a6fd5ea9adf0631f9ade27f6c
import torch import torch.nn as nn class Model(nn.Module): """ Expects :param x: [examples, samples, channels, H, W] :return: [examples, samples, channels] """ def __init__(self): super().__init__() def forward(self, x): x = x.mean(4).mean(3) return x def get_in...
MAPE
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class MAPE(nn.Module): def __init__(self): super().__init__() self.l1 = nn.L1Loss(reduction='none') def forward(self, pred, actual): mape = 100 * self.l1(pred, actual) / torch.max(pred, actual) return mape.mean() def get_inputs(): retu...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
RosarioAndolina/psychXRF
MAPE
false
990
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() self.l1 = nn.L1Loss(reduction='none') def forward(self, pred, actual): mape = 100 * self.l1(pred, actual) / torch.max(pred, actual) return mape.mean() def get_inputs(): ret...
MultiheadAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter class MultiheadAttention(nn.Module): """Multi-headed attention. See "Attention Is All You Need" for more details. """ def __init__(self, embed_dim, num_heads, attn_dropout=0.0, bias=True, add_bia...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Redaimao/RERD
MultiheadAttention
false
991
[ "MIT" ]
0
40413d4b6743f3e5db0c30ee969d45661d001834
https://github.com/Redaimao/RERD/tree/40413d4b6743f3e5db0c30ee969d45661d001834
import torch import torch.nn as nn import torch.nn.functional as F from torch.nn import Parameter class Model(nn.Module): """Multi-headed attention. See "Attention Is All You Need" for more details. """ def __init__(self, embed_dim, num_heads, attn_dropout=0.0, bias=True, add_bias_kv=False, a...
SVIGlobalMaxPool2D
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class SVIGlobalMaxPool2D(nn.Module): """ Expects :param x: [examples, samples, channels, H, W] :return: [examples, samples, channels] """ def __init__(self): super(SVIGlobalMaxPool2D, self).__init__() def forward(self, x): x = x.max(4)[0...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
RomanShen/radial-bnn
SVIGlobalMaxPool2D
false
992
[ "Apache-2.0" ]
0
7c8bc85397c1461a6fd5ea9adf0631f9ade27f6c
https://github.com/RomanShen/radial-bnn/tree/7c8bc85397c1461a6fd5ea9adf0631f9ade27f6c
import torch import torch.nn as nn class Model(nn.Module): """ Expects :param x: [examples, samples, channels, H, W] :return: [examples, samples, channels] """ def __init__(self): super().__init__() def forward(self, x): x = x.max(4)[0].max(3)[0] return x def ge...
Critic
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn.functional as f from torch import nn class Critic(nn.Module): def __init__(self, input_dim): super(Critic, self).__init__() self._input_dim = input_dim self.dense1 = nn.Linear(self._input_dim, self._input_dim) self.dense2 = nn.Linear(self._input_dim, s...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch import nn assert_size_stride = torch._C._dynamo.guards.assert_size_st...
RosalRicardo/RTRGAN
Critic
false
993
[ "MIT" ]
0
6f4551ab8445367f8b9c711f41f15dd465abaef1
https://github.com/RosalRicardo/RTRGAN/tree/6f4551ab8445367f8b9c711f41f15dd465abaef1
import torch import torch.nn.functional as f from torch import nn class Model(nn.Module): def __init__(self, input_dim): super().__init__() self._input_dim = input_dim self.dense1 = nn.Linear(self._input_dim, self._input_dim) self.dense2 = nn.Linear(self._input_dim, self._input_di...
R2Score
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class R2Score(nn.Module): def __init__(self): super().__init__() def forward(self, pred, actual): rss = ((actual - pred) ** 2).sum() ym = actual.mean() tss = ((actual - ym) ** 2).sum() return 1 - rss / tss def get_inputs(): ret...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
RosarioAndolina/psychXRF
R2Score
false
994
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, pred, actual): rss = ((actual - pred) ** 2).sum() ym = actual.mean() tss = ((actual - ym) ** 2).sum() return 1 - rss / tss def get_inputs(): retur...
MultiHeadAttention
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn as nn import torch.nn.functional as F class MultiHeadAttention(nn.Module): def __init__(self, d_model, n_heads, p_drop=0.1): super(MultiHeadAttention, self).__init__() assert d_model % n_heads == 0 self.d_model = d_model self.n_heads...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
RegiusQuant/nlp-practice
MultiHeadAttention
false
995
[ "MIT" ]
0
ffa99aa585134941aa148da11775c2b16d42eef7
https://github.com/RegiusQuant/nlp-practice/tree/ffa99aa585134941aa148da11775c2b16d42eef7
import torch import numpy as np import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, d_model, n_heads, p_drop=0.1): super().__init__() assert d_model % n_heads == 0 self.d_model = d_model self.n_heads = n_heads self.d_hidden = d_...
ACNet
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import numpy as np import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable def from_numpy(np_array, dtype=np.float32): if np_array.dtype != dtype: np_array = np_array.astype(dtype) return Variable(torch.from_numpy(np_array)) class ACNet(nn.Module): ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import numpy as np import tor...
RocksonZeta/ac
ACNet
false
996
[ "MIT" ]
0
050a5cd176864cc2e1f7c376045c3342a7f93221
https://github.com/RocksonZeta/ac/tree/050a5cd176864cc2e1f7c376045c3342a7f93221
import torch import numpy as np import torch.nn as nn import torch.nn.functional as F from torch.autograd import Variable def from_numpy(np_array, dtype=np.float32): if np_array.dtype != dtype: np_array = np_array.astype(dtype) return Variable(torch.from_numpy(np_array)) class Model(nn.Module): ...
BReLU
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class BReLU(nn.Module): """ Biased ReLU BReLU(x) = ReLU(x) + b Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -in_features: number of input features -b: fixed parameter (bias like for relu...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
RosarioAndolina/psychXRF
BReLU
false
997
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): """ Biased ReLU BReLU(x) = ReLU(x) + b Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -in_features: number of input features -b: fixed parameter (bias like for relu...
Beta2
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import numpy as np import torch.nn as nn class BoundedBeta(torch.distributions.Beta): def log_prob(self, x): return super().log_prob((x + 1) / 2) class Beta2(nn.Module): def __init__(self, action_dim, init_std=0.25, learn_std=False): super(Beta2, self).__init__() asser...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import numpy as np import torch.nn as nn assert_size_stride = torch._C._d...
RohanPankaj/apex
Beta2
false
998
[ "MIT" ]
0
74e96386bf9446d1179106d6d65ea0368c1b5b27
https://github.com/RohanPankaj/apex/tree/74e96386bf9446d1179106d6d65ea0368c1b5b27
import torch import numpy as np import torch.nn as nn class BoundedBeta(torch.distributions.Beta): def log_prob(self, x): return super().log_prob((x + 1) / 2) class Model(nn.Module): def __init__(self, action_dim, init_std=0.25, learn_std=False): super().__init__() assert init_std ...
LN_Actor
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class LN_Actor(nn.Module): def __init__(self, state_dim, action_dim, max_action, hidden_size1, hidden_size2): super(LN_Actor, self).__init__() self.l1 = nn.Linear(state_dim, hidden_size1) self.ln1 = nn.LayerNorm(hi...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
RohanPankaj/apex
LN_Actor
false
999
[ "MIT" ]
0
74e96386bf9446d1179106d6d65ea0368c1b5b27
https://github.com/RohanPankaj/apex/tree/74e96386bf9446d1179106d6d65ea0368c1b5b27
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, state_dim, action_dim, max_action, hidden_size1, hidden_size2): super().__init__() self.l1 = nn.Linear(state_dim, hidden_size1) self.ln1 = nn.LayerNorm(hidden_size1) ...
CFReLU
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class CFReLU(nn.Module): """ Custom FReLU cfrelu(x) = relu(x + a) + b see psychXRF.activation.FReLU Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -a: trai...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn import torch.nn.functional as F assert_size_stride = torch._C._dyna...
RosarioAndolina/psychXRF
CFReLU
false
1,000
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): """ Custom FReLU cfrelu(x) = relu(x + a) + b see psychXRF.activation.FReLU Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -a: train...
LN_DDPGCritic
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class LN_DDPGCritic(nn.Module): def __init__(self, state_dim, action_dim, hidden_size1, hidden_size2): super(LN_DDPGCritic, self).__init__() self.l1 = nn.Linear(state_dim + action_dim, hidden_size1) self.ln1 = nn.LayerNorm...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
RohanPankaj/apex
LN_DDPGCritic
false
1,001
[ "MIT" ]
0
74e96386bf9446d1179106d6d65ea0368c1b5b27
https://github.com/RohanPankaj/apex/tree/74e96386bf9446d1179106d6d65ea0368c1b5b27
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, state_dim, action_dim, hidden_size1, hidden_size2): super().__init__() self.l1 = nn.Linear(state_dim + action_dim, hidden_size1) self.ln1 = nn.LayerNorm(hidden_size1) self...
ReQUNet
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn def MyReQU(x): x[x < 0] = 0 z = x * x return z class ReQUNet(nn.Module): def __init__(self): super(ReQUNet, self).__init__() n_in, n_h, n_out = 4, 64, 3 self.fc1 = nn.Linear(n_in, n_h, True) self.fc2 = nn.Linear(n_h, n_out, True) ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
RoyHirsch/DeepLearningCourse
ReQUNet
false
1,002
[ "MIT" ]
0
9036c0fdbb08b610524d7be991f8e4b490a82c6c
https://github.com/RoyHirsch/DeepLearningCourse/tree/9036c0fdbb08b610524d7be991f8e4b490a82c6c
import torch import torch.nn as nn def MyReQU(x): x[x < 0] = 0 z = x * x return z class Model(nn.Module): def __init__(self): super().__init__() n_in, n_h, n_out = 4, 64, 3 self.fc1 = nn.Linear(n_in, n_h, True) self.fc2 = nn.Linear(n_h, n_out, True) def forward(...
SumNorm
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class SumNorm(nn.Module): """ Normalize dividing by the sum Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -in_features: number of input features -dim(int): A dimension along witch sum will be comp...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
RosarioAndolina/psychXRF
SumNorm
false
1,003
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): """ Normalize dividing by the sum Shape: -Input: (N, *) -Output: (N, *), same shape as the input Parameters: -in_features: number of input features -dim(int): A dimension along witch sum will be comput...
PITF_Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch as t import torch.nn as nn class PITF_Loss(nn.Module): """ 定义PITF的loss function """ def __init__(self): super(PITF_Loss, self).__init__() None def forward(self, r_p, r_ne): return -t.log(t.sigmoid(r_p - r_ne)) def get_inputs(): return [torc...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert...
SamHaoYuan/pitf
PITF_Loss
false
1,004
[ "MIT" ]
0
5fdebc3b44c6462126876101b052a3980804da79
https://github.com/SamHaoYuan/pitf/tree/5fdebc3b44c6462126876101b052a3980804da79
import torch import torch as t import torch.nn as nn class Model(nn.Module): """ 定义PITF的loss function """ def __init__(self): super().__init__() None def forward(self, r_p, r_ne): return -t.log(t.sigmoid(r_p - r_ne)) def get_inputs(): return [torch.rand([4, 4, 4, 4]...
SpatialGather_Module
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn from torch.nn import functional as F import torch._utils import torch.optim class SpatialGather_Module(nn.Module): """ Aggregate the context features according to the initial predicted probability distribution. Employ the soft-weighted method to aggregat...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
SSJIACV/HRNet-Semantic-Segmentation
SpatialGather_Module
false
1,005
[ "MIT" ]
0
7e2840ce7a91ae3845dfb203c992f84affa15e40
https://github.com/SSJIACV/HRNet-Semantic-Segmentation/tree/7e2840ce7a91ae3845dfb203c992f84affa15e40
import torch import torch.nn as nn from torch.nn import functional as F import torch._utils import torch.optim class Model(nn.Module): """ Aggregate the context features according to the initial predicted probability distribution. Employ the soft-weighted method to aggregate the context. ...
_boundary
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn class _boundary(nn.Module): def __init__(self, dim): super(_boundary, self).__init__() self.relu = nn.ReLU(inplace=True) self.conv1 = nn.Conv2d(dim, dim, kernel_size=3, padding=1) self.conv2 = nn.Conv2d(dim, dim, kernel_size=3, padding=1) def...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn assert_s...
STARBOYsachin/semantic-segmentation
_boundary
false
1,006
[ "MIT" ]
0
7f553a93b717641edc6c2d463903dfab67267039
https://github.com/STARBOYsachin/semantic-segmentation/tree/7f553a93b717641edc6c2d463903dfab67267039
import torch from torch import nn class Model(nn.Module): def __init__(self, dim): super().__init__() self.relu = nn.ReLU(inplace=True) self.conv1 = nn.Conv2d(dim, dim, kernel_size=3, padding=1) self.conv2 = nn.Conv2d(dim, dim, kernel_size=3, padding=1) def forward(self, x): ...
SinglePITF_Loss
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch as t import torch.nn as nn class SinglePITF_Loss(nn.Module): """ 定义PITF的loss function """ def __init__(self): super(SinglePITF_Loss, self).__init__() None def forward(self, r): return t.sum(-t.log(t.sigmoid(r))) def get_inputs(): return [to...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn ...
SamHaoYuan/pitf
SinglePITF_Loss
false
1,007
[ "MIT" ]
0
5fdebc3b44c6462126876101b052a3980804da79
https://github.com/SamHaoYuan/pitf/tree/5fdebc3b44c6462126876101b052a3980804da79
import torch import torch as t import torch.nn as nn class Model(nn.Module): """ 定义PITF的loss function """ def __init__(self): super().__init__() None def forward(self, r): return t.sum(-t.log(t.sigmoid(r))) def get_inputs(): return [torch.rand([4, 4, 4, 4])] def g...
CAMMNISTClassifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn as nn from torch import optim as optim from torchvision import transforms as transforms class CAMMNISTClassifier(nn.Module): def __init__(self): super(CAMMNISTClassifier, self).__init__() self.conv1 = nn.Conv2d(1, 32, kernel_size=3) self.relu1 = nn.ReLU()...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn as nn fr...
RobinMaas95/GTSRB_Visualization
CAMMNISTClassifier
false
1,008
[ "MIT" ]
0
fa837ff94e089a936ef4f4418970d262b35f70b6
https://github.com/RobinMaas95/GTSRB_Visualization/tree/fa837ff94e089a936ef4f4418970d262b35f70b6
import torch from torch import nn as nn from torch import optim as optim from torchvision import transforms as transforms class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(1, 32, kernel_size=3) self.relu1 = nn.ReLU() self.pool1 = nn.MaxPool2d(2)...
Standardize
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import torch from torch.nn import init from torch.nn.parameter import Parameter class Standardize(Module): """ Applies (element-wise) standardization with trainable translation parameter μ and scale parameter σ, i.e. computes (x - μ) / σ where '/' is applied element-wise. ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch.nn import Module from torch.nn import init from torch.nn.parameter import Parameter assert_size_stride = torch._C._dynamo.guards....
SDJustus/Deep-SAD-PyTorch
Standardize
false
1,009
[ "MIT" ]
0
4d98e6474a7256329134c075894f885a56f59281
https://github.com/SDJustus/Deep-SAD-PyTorch/tree/4d98e6474a7256329134c075894f885a56f59281
from torch.nn import Module import torch from torch.nn import init from torch.nn.parameter import Parameter class Model(Module): """ Applies (element-wise) standardization with trainable translation parameter μ and scale parameter σ, i.e. computes (x - μ) / σ where '/' is applied element-wise. Args: ...
NetFCN12
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class NetFCN12(nn.Module): def __init__(self): super(NetFCN12, self).__init__() self.conv = nn.Conv2d(3, 16, 3) self.pool = nn.MaxPool2d((3, 3), stride=2) self.conv2 = nn.Conv2d(16, 16, 4) self.conv3 = nn.C...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
RoyHirsch/DeepLearningCourse
NetFCN12
false
1,010
[ "MIT" ]
0
9036c0fdbb08b610524d7be991f8e4b490a82c6c
https://github.com/RoyHirsch/DeepLearningCourse/tree/9036c0fdbb08b610524d7be991f8e4b490a82c6c
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self): super().__init__() self.conv = nn.Conv2d(3, 16, 3) self.pool = nn.MaxPool2d((3, 3), stride=2) self.conv2 = nn.Conv2d(16, 16, 4) self.conv3 = nn.Conv2d(16, 2, 1) ...
AttentionConv
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class AttentionConv(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, bias=False): super(AttentionConv, self).__init__() self.out_channels = ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch....
Sam-limyr/End-to-end-ASR-Pytorch
AttentionConv
false
1,011
[ "MIT" ]
0
623a50792f48218228549ea17b8ea5e8bb1b342f
https://github.com/Sam-limyr/End-to-end-ASR-Pytorch/tree/623a50792f48218228549ea17b8ea5e8bb1b342f
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class Model(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, bias=False): super().__init__() self.out_channels = out_channels self.k...
ResBlock
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class ResBlock(nn.Module): def __init__(self, size): super().__init__() self.size = size self.layer1 = nn.Linear(self.size, self.size) self.relu = nn.ReLU() self.layer2 = nn.Linear(self.size, self.size) def forward(self, x): ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
RosarioAndolina/psychXRF
ResBlock
false
1,012
[ "MIT" ]
0
e2adadbd17664d7f74c10304f84b3751c571226e
https://github.com/RosarioAndolina/psychXRF/tree/e2adadbd17664d7f74c10304f84b3751c571226e
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, size): super().__init__() self.size = size self.layer1 = nn.Linear(self.size, self.size) self.relu = nn.ReLU() self.layer2 = nn.Linear(self.size, self.size) def forward(self, x): sho...
LN_TD3Critic
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class LN_TD3Critic(nn.Module): def __init__(self, state_dim, action_dim, hidden_size1, hidden_size2): super(LN_TD3Critic, self).__init__() self.l1 = nn.Linear(state_dim + action_dim, hidden_size1) self.ln1 = nn.LayerNorm(h...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
RohanPankaj/apex
LN_TD3Critic
false
1,013
[ "MIT" ]
0
74e96386bf9446d1179106d6d65ea0368c1b5b27
https://github.com/RohanPankaj/apex/tree/74e96386bf9446d1179106d6d65ea0368c1b5b27
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, state_dim, action_dim, hidden_size1, hidden_size2): super().__init__() self.l1 = nn.Linear(state_dim + action_dim, hidden_size1) self.ln1 = nn.LayerNorm(hidden_size1) self...
Transformer
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch as t import torch.nn as nn from torch.distributions.categorical import Categorical from torch.autograd import Variable import torch.nn.functional as F import torch.optim as optim class Transformer(nn.Module): def __init__(self, input_size, num_actions, hidden_size, learning_rate= ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
LucWeber/2048-RLenv
Transformer
false
1,014
[ "MIT" ]
0
6beff54691f0436f0fbca6bdbb9430fd37eab37d
https://github.com/LucWeber/2048-RLenv/tree/6beff54691f0436f0fbca6bdbb9430fd37eab37d
import torch import torch as t import torch.nn as nn from torch.distributions.categorical import Categorical from torch.autograd import Variable import torch.nn.functional as F import torch.optim as optim class Model(nn.Module): def __init__(self, input_size, num_actions, hidden_size, learning_rate= 0.00...
TwoLayerNet
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class TwoLayerNet(torch.nn.Module): def __init__(self, D_in, H, D_out): super(TwoLayerNet, self).__init__() self.linear1 = torch.nn.Linear(D_in, H) self.linear2 = torch.nn.Linear(H, D_out) def forward(self, x): h_relu = self.linear1(x).clamp(min=0) y_pred...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers assert_size_stride = torch._C...
Saran-nns/delve
TwoLayerNet
false
1,015
[ "MIT" ]
0
3489d8aa13181b392d3c47a19f9d9a47d87f8790
https://github.com/Saran-nns/delve/tree/3489d8aa13181b392d3c47a19f9d9a47d87f8790
import torch class Model(torch.nn.Module): def __init__(self, D_in, H, D_out): super().__init__() self.linear1 = torch.nn.Linear(D_in, H) self.linear2 = torch.nn.Linear(H, D_out) def forward(self, x): h_relu = self.linear1(x).clamp(min=0) y_pred = self.linear2(h_relu)...
PixelwiseNorm
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch class PixelwiseNorm(torch.nn.Module): """ ------------------------------------------------------------------------------------ Pixelwise feature vector normalization. reference: https://github.com/tkarras/progressive_growing_of_gans/blob/master/networks.py#L120 -------------------...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_c...
SashaMatsun/torch-GAN
PixelwiseNorm
false
1,016
[ "MIT" ]
0
534a634530548d3f8b3a102c3e43e1cc64d8506d
https://github.com/SashaMatsun/torch-GAN/tree/534a634530548d3f8b3a102c3e43e1cc64d8506d
import torch class Model(torch.nn.Module): """ ------------------------------------------------------------------------------------ Pixelwise feature vector normalization. reference: https://github.com/tkarras/progressive_growing_of_gans/blob/master/networks.py#L120 ---------------------------...
CAMMNISTExtendedClassifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch from torch import nn as nn from torch import optim as optim from torchvision import transforms as transforms from torch.nn.functional import pad class CAMMNISTExtendedClassifier(nn.Module): def __init__(self): super(CAMMNISTExtendedClassifier, self).__init__() self.conv1 = nn.Conv2d(...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch import nn as nn fr...
RobinMaas95/GTSRB_Visualization
CAMMNISTExtendedClassifier
false
1,017
[ "MIT" ]
0
fa837ff94e089a936ef4f4418970d262b35f70b6
https://github.com/RobinMaas95/GTSRB_Visualization/tree/fa837ff94e089a936ef4f4418970d262b35f70b6
import torch from torch import nn as nn from torch import optim as optim from torchvision import transforms as transforms from torch.nn.functional import pad class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d(1, 32, kernel_size=4) self.relu1 = nn.ReLU() ...
Model
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class Model(torch.nn.Module): def __init__(self, D_in, D_out): super(Model, self).__init__() self.w1 = torch.nn.Parameter(torch.randn(D_in, D_out), requires_grad=True) self.sig = torch.nn.Sigmoid() def forward(self, x): y_pred = self.sig(x.mm(self.w1)...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cu...
Saran-nns/gradients
Model
false
1,018
[ "MIT" ]
0
67f9ff92589047828563dbbe30f225dca7ad47fd
https://github.com/Saran-nns/gradients/tree/67f9ff92589047828563dbbe30f225dca7ad47fd
import torch class Model(torch.nn.Module): def __init__(self, D_in, D_out): super(Model, self).__init__() self.w1 = torch.nn.Parameter(torch.randn(D_in, D_out), requires_grad=True) self.sig = torch.nn.Sigmoid() def forward(self, x): y_pred = self.sig(x.mm(self.w1)...
DisentangledAELatent
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class DisentangledAELatent(torch.nn.Module): """Dense Dientangled Latent Layer between encoder and decoder""" def __init__(self, hidden_size: 'int', latent_size: 'int', dropout: 'float' ): super(DisentangledAELatent, self).__init__() self.latent_size = latent_size ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math assert_size_s...
Saran-nns/traja
DisentangledAELatent
false
1,019
[ "MIT" ]
0
f2256cc47abd33377b3a87f110f4c8da1cf6765f
https://github.com/Saran-nns/traja/tree/f2256cc47abd33377b3a87f110f4c8da1cf6765f
import torch class Model(torch.nn.Module): """Dense Dientangled Latent Layer between encoder and decoder""" def __init__(self, hidden_size: 'int', latent_size: 'int', dropout: 'float' ): super().__init__() self.latent_size = latent_size self.hidden_size = hidden_size s...
LayerCake
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch class LayerCake(torch.nn.Module): def __init__(self, D_in, H1, H2, H3, H4, H5, D_out): """ In the constructor we instantiate two nn.Linear modules and assign them as member variables. """ super(LayerCake, self).__init__() self.linear1 = torch.nn.Linear...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers assert_size_stride = torch._C...
Saran-nns/delve
LayerCake
false
1,021
[ "MIT" ]
0
3489d8aa13181b392d3c47a19f9d9a47d87f8790
https://github.com/Saran-nns/delve/tree/3489d8aa13181b392d3c47a19f9d9a47d87f8790
import torch class Model(torch.nn.Module): def __init__(self, D_in, H1, H2, H3, H4, H5, D_out): """ In the constructor we instantiate two nn.Linear modules and assign them as member variables. """ super().__init__() self.linear1 = torch.nn.Linear(D_in, H1) ...
GCN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
from torch.nn import Module import math import torch import torch.nn as nn import torch.utils.data import torch from torch.nn.modules.module import Module import torch.nn.functional as F from torch.nn.parameter import Parameter class GCN_Spectral(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/160...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
SamujjwalSam/XC_GCN
GCN
false
1,022
[ "MIT" ]
0
7902cbd6b3ebc7806655080979e8c52caa4a16e0
https://github.com/SamujjwalSam/XC_GCN/tree/7902cbd6b3ebc7806655080979e8c52caa4a16e0
from torch.nn import Module import math import torch import torch.nn as nn import torch.utils.data import torch from torch.nn.modules.module import Module import torch.nn.functional as F from torch.nn.parameter import Parameter class GCN_Spectral(Module): """ Simple GCN layer, similar to https://arxiv.org/abs/160...
ReOrgLayer
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn import torch.utils.data import torch.utils.data.distributed import torch._utils class ReOrgLayer(nn.Module): def __init__(self, stride=2): super(ReOrgLayer, self).__init__() self.stride = stride def forward(self, x): assert x.data.dim() == 4 ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn import torch.utils.data import torch.utils.data.distributed import torch._utils assert_size_stride = torch._C._dynamo....
Sarathismg/Pose-Estimator-Old-Version
ReOrgLayer
false
1,023
[ "Apache-2.0" ]
0
ecaa03769323b94a4d7222e2d3606d1ce92a2fae
https://github.com/Sarathismg/Pose-Estimator-Old-Version/tree/ecaa03769323b94a4d7222e2d3606d1ce92a2fae
import torch import torch.nn as nn import torch.utils.data import torch.utils.data.distributed import torch._utils class Model(nn.Module): def __init__(self, stride=2): super().__init__() self.stride = stride def forward(self, x): assert x.data.dim() == 4 B, C, H, W = x.data....
TracedModule
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.onnx import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed import torch.autograd class TracedModule(torch.nn.Module): def forward(self, x): x = x.type(torch.float32) return torch.floor(torch.sqrt(x) / 5.0) def get_input...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import libdevice import torch.onnx import torch.nn.parallel import torch.optim import torch.util...
ScorpioDoctor/antares02
TracedModule
false
1,024
[ "BSD-3-Clause" ]
0
631b817d2e98f351d1173b620d15c4a5efed11da
https://github.com/ScorpioDoctor/antares02/tree/631b817d2e98f351d1173b620d15c4a5efed11da
import torch import torch.onnx import torch.nn.parallel import torch.optim import torch.utils.data import torch.utils.data.distributed import torch.autograd class Model(torch.nn.Module): def forward(self, x): x = x.type(torch.float32) return torch.floor(torch.sqrt(x) / 5.0) def get_inputs(): ...
SRCNN
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import logging import torch import torchvision import warnings from collections import OrderedDict from torch.utils import model_zoo from torch.nn import functional as F import torch.nn as nn def get_root_logger(log_file=None, log_level=logging.INFO): """Get the root logger. The logger will be initialized if...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Jason-Khan/mmediting
SRCNN
false
1,025
[ "Apache-2.0" ]
0
d187f95a675dff3eb975a575bd9278d643b5b645
https://github.com/Jason-Khan/mmediting/tree/d187f95a675dff3eb975a575bd9278d643b5b645
import logging import torch import torchvision import warnings from collections import OrderedDict from torch.utils import model_zoo from torch.nn import functional as F import torch.nn as nn def get_root_logger(log_file=None, log_level=logging.INFO): """Get the root logger. The logger will be initialized if...
PairCosineSim
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.utils.data import torch.nn as nn class PairCosineSim(nn.Module): def __init__(self): super(PairCosineSim, self).__init__() def forward(self, supports, target): """ Calculates pairwise cosine similarity of support sets with target sample. :param supp...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
SamujjwalSam/MatchingNetworks4XC
PairCosineSim
false
1,026
[ "MIT" ]
0
2519cc1a527ea121c4966c1a860d890d5182f887
https://github.com/SamujjwalSam/MatchingNetworks4XC/tree/2519cc1a527ea121c4966c1a860d890d5182f887
import torch import torch.utils.data import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, supports, target): """ Calculates pairwise cosine similarity of support sets with target sample. :param supports: The embeddings of the...
squeeze
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class squeeze(nn.Module): def __init__(self, block_size): super(squeeze, self).__init__() self.block_size = block_size self.block_size_sq = block_size * block_size def inverse(self, input): output = input.permute(0, 2, 3, 1) batch_si...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Schwartz-Zha/My-invertible-resnet
squeeze
false
1,027
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, block_size): super().__init__() self.block_size = block_size self.block_size_sq = block_size * block_size def inverse(self, input): output = input.permute(0, 2, 3, 1) batch_size, d_height, d...
MaxMinGroup
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn def process_maxmin_groupsize(x, group_size, axis=-1): size = list(x.size()) num_channels = size[axis] if num_channels % group_size: raise ValueError( 'number of features({}) is not a multiple of group_size({})'. format(num_channels, num_ch...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride emp...
Schwartz-Zha/My-invertible-resnet
MaxMinGroup
false
1,028
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn def process_maxmin_groupsize(x, group_size, axis=-1): size = list(x.size()) num_channels = size[axis] if num_channels % group_size: raise ValueError( 'number of features({}) is not a multiple of group_size({})'. format(num_channels, num_ch...
Split
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class Split(nn.Module): def __init__(self): super(Split, self).__init__() def forward(self, x): n = int(x.size(1) / 2) x1 = x[:, :n, :, :].contiguous() x2 = x[:, n:, :, :].contiguous() return x1, x2 def inverse(self, x1, x2): ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Schwartz-Zha/My-invertible-resnet
Split
false
1,029
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Module): def __init__(self): super().__init__() def forward(self, x): n = int(x.size(1) / 2) x1 = x[:, :n, :, :].contiguous() x2 = x[:, n:, :, :].contiguous() return x1, x2 def inverse(self, x1, x2): retur...
NormalAttention_gaussian
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class NormalAttention_gaussian(nn.Module): def __init__(self, input_channel_num): super(NormalAttention_gaussian, self).__init__() self.c_in = input_channel_num self.value_conv = nn.Conv2d(in_channels=self.c_in, out_channels= self.c_in, kerne...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch....
Schwartz-Zha/My-invertible-resnet
NormalAttention_gaussian
false
1,030
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, input_channel_num): super().__init__() self.c_in = input_channel_num self.value_conv = nn.Conv2d(in_channels=self.c_in, out_channels= self.c_in, kernel_size=1) self.gamma = nn.Conv2d(in_chann...
Conv2dZeroInit
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class Conv2dZeroInit(nn.Conv2d): def __init__(self, channels_in, channels_out, filter_size, stride=1, padding=0, logscale=3.0): super().__init__(channels_in, channels_out, filter_size, stride= stride, padding=padding) self.register_parameter(...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch....
Schwartz-Zha/My-invertible-resnet
Conv2dZeroInit
false
1,031
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Conv2d): def __init__(self, channels_in, channels_out, filter_size, stride=1, padding=0, logscale=3.0): super().__init__(channels_in, channels_out, filter_size, stride= stride, padding=padding) self.register_parameter('logs', n...
Net
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F class Net(nn.Module): def __init__(self, h2): super(Net, self).__init__() self.conv1 = nn.Conv2d(3, 6, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 = nn.Conv2d(6, 16, 5) self.fc1 = nn.Linear(16 * 5 * 5, 120...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers import torch.nn as nn assert_...
Saran-nns/delve
Net
false
1,032
[ "MIT" ]
0
3489d8aa13181b392d3c47a19f9d9a47d87f8790
https://github.com/Saran-nns/delve/tree/3489d8aa13181b392d3c47a19f9d9a47d87f8790
import torch import torch.nn as nn import torch.nn.functional as F class Model(nn.Module): def __init__(self, h2): super().__init__() self.conv1 = nn.Conv2d(3, 6, 5) self.pool = nn.MaxPool2d(2, 2) self.conv2 = nn.Conv2d(6, 16, 5) self.fc1 = nn.Linear(16 * 5 * 5, 120) ...
MeanVarFC
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn class MeanVarFC(nn.Module): def __init__(self, input_shape): super(MeanVarFC, self).__init__() shape = list(input_shape) shape[0] = 1 shape[1] *= 2 self.param = nn.Parameter(0.01 * torch.randn(shape)) def forward(self, x): x ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Schwartz-Zha/My-invertible-resnet
MeanVarFC
false
1,033
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, input_shape): super().__init__() shape = list(input_shape) shape[0] = 1 shape[1] *= 2 self.param = nn.Parameter(0.01 * torch.randn(shape)) def forward(self, x): x = x + self.param ...
injective_pad
# AOT ID: ['0_inference'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _al...
import torch import torch.nn as nn class injective_pad(nn.Module): def __init__(self, pad_size): super(injective_pad, self).__init__() self.pad_size = pad_size self.pad = nn.ZeroPad2d((0, 0, 0, pad_size)) def forward(self, x): x = x.permute(0, 2, 1, 3) x = self.pad(x)...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream import torch.nn as nn assert_size_stride = torch._C._dynamo.guards.assert_size_stride empty_strided_cuda = torch._C._dynamo.guards._empty_st...
Schwartz-Zha/My-invertible-resnet
injective_pad
false
1,034
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn class Model(nn.Module): def __init__(self, pad_size): super().__init__() self.pad_size = pad_size self.pad = nn.ZeroPad2d((0, 0, 0, pad_size)) def forward(self, x): x = x.permute(0, 2, 1, 3) x = self.pad(x) return x.permute(0...
ANNClassifier
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class AttentionConv(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, bias=False): super(AttentionConv, self).__init__() self.out_channels = ...
import torch from torch._inductor.select_algorithm import extern_kernels import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime import triton_helpers from torch._inductor.runtime....
Sam-limyr/End-to-end-ASR-Pytorch
ANNClassifier
false
1,035
[ "MIT" ]
0
623a50792f48218228549ea17b8ea5e8bb1b342f
https://github.com/Sam-limyr/End-to-end-ASR-Pytorch/tree/623a50792f48218228549ea17b8ea5e8bb1b342f
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class AttentionConv(nn.Module): def __init__(self, in_channels, out_channels, kernel_size, stride=1, padding=0, groups=1, bias=False): super().__init__() self.out_channels = out_channels ...
ActNorm
# AOT ID: ['0_forward'] from ctypes import c_void_p, c_long, c_int import torch import math import random import os import tempfile from math import inf, nan from torch._inductor.hooks import run_intermediate_hooks from torch._inductor.utils import maybe_profile from torch._inductor.codegen.memory_planning import _alig...
import torch import torch.nn as nn from torch.nn import Parameter class ActNorm(nn.Module): def __init__(self, num_channels, eps=1e-05): super(ActNorm, self).__init__() self.eps = eps self.num_channels = num_channels self._log_scale = Parameter(torch.Tensor(num_channels)) ...
import torch import triton import triton.language as tl from torch._inductor.runtime.triton_heuristics import grid from torch._C import _cuda_getCurrentRawStream as get_raw_stream from torch._inductor.runtime.triton_helpers import math as tl_math import torch.nn as nn from torch.nn import Parameter assert_size_stride =...
Schwartz-Zha/My-invertible-resnet
ActNorm
false
1,036
[ "MIT" ]
0
5415975bb0d640f3bf3ef4a7b986563e84109270
https://github.com/Schwartz-Zha/My-invertible-resnet/tree/5415975bb0d640f3bf3ef4a7b986563e84109270
import torch import torch.nn as nn from torch.nn import Parameter class Model(nn.Module): def __init__(self, num_channels, eps=1e-05): super().__init__() self.eps = eps self.num_channels = num_channels self._log_scale = Parameter(torch.Tensor(num_channels)) self._shift = P...