text stringlengths 2.5k 6.39M | kind stringclasses 3
values |
|---|---|
#include "dragon/core/context_cuda.h"
#include "dragon/utils/math_functions.h"
#include "dragon/utils/op_kernels.h"
namespace dragon {
namespace kernels {
namespace {
#define LDG(x, i) __ldg(x + i)
#define LDG2(x, i) convert::To<float>(__ldg(x + i))
template <typename T>
__global__ void _ResizeNearest2dNCHW(
c... | the_stack |
#include<ops/declarable/helpers/sru.h>
#include <array/NDArrayFactory.h>
#include <helpers/PointersManager.h>
#include <helpers/MmulHelper.h>
namespace sd {
namespace ops {
namespace helpers {
//////////////////////////////////////////////////////////////////////////
static FORCEINLINE NDArray activati... | the_stack |
// -----------------------------------------------------------------------------------------
// NVEnc by rigaya
// -----------------------------------------------------------------------------------------
//
// The MIT License
//
// Copyright (c) 2014-2016 rigaya
//
// Permission is hereby granted, free of cha... | the_stack |
* \file
* cub::AgentSelectIf implements a stateful abstraction of CUDA thread blocks for participating in device-wide select.
*/
#pragma once
#include <iterator>
#include "single_pass_scan_operators.cuh"
#include "../block/block_load.cuh"
#include "../block/block_store.cuh"
#include "../block/block_scan.cuh"
#incl... | the_stack |
#include "cupoch/geometry/pointcloud.h"
#include "cupoch/integration/integrate_functor.h"
#include "cupoch/integration/marching_cubes_const.h"
#include "cupoch/integration/scalable_tsdfvolume.h"
#include "cupoch/utility/console.h"
#include "cupoch/utility/platform.h"
#include "cupoch/utility/range.h"
namespace cupoch ... | the_stack |
//Includes for IntelliSense
#define _SIZE_T_DEFINED
#ifndef __CUDACC__
#define __CUDACC__
#endif
#ifndef __cplusplus
#define __cplusplus
#endif
extern "C"
{
//kernel code
__global__ void VectorInputDiffKernel(
float *input,
int inputSize,
float *referenceVector,
int maxCells,
float *difference
... | the_stack |
#include "anonymouslib_cuda.h"
#include "mmio.h"
using namespace std;
#ifndef VALUE_TYPE
#define VALUE_TYPE double
#endif
#ifndef NUM_RUN
#define NUM_RUN 1000
#endif
int call_anonymouslib(int m, int n, int nnzA,
int *csrRowPtrA, int *csrColIdxA, VALUE_TYPE *csrValA,
VALUE_TYPE *... | the_stack |
#include "grid_sample_kernel_util.h"
namespace oneflow {
class CudnnGridSampleDesc final {
public:
OF_DISALLOW_COPY_AND_MOVE(CudnnGridSampleDesc);
CudnnGridSampleDesc(DataType data_type, const ShapeView& shape) {
std::vector<int> tensor_dim({shape.ptr(), shape.ptr() + shape.NumAxes()});
OF_CUDNN_CHECK(cu... | the_stack |
#ifndef IDIVUP
#define IDIVUP(i,j) ((i+j-1)/j)
#endif
///
/// Constructor for cuAmpcorParameter class
/// also sets the default/initial values of various parameters
///
cuAmpcorParameter::cuAmpcorParameter()
{
// default settings
// will be changed if they are set by python scripts
algorithm = 0; //0 freq... | the_stack |
#include <THC/THCDeviceUtils.cuh>
#include <THC/THCGeneral.h>
#include <ATen/ATen.h>
#include <ATen/AccumulateType.h>
#include <ATen/cuda/CUDAContext.h>
#include <ATen/cuda/CUDAApplyUtils.cuh>
#include <ATen/TensorAccessor.h>
#if defined(__HIP_PLATFORM_HCC__)
constexpr int WARP_SIZE = 64;
#else
constexpr int WARP_SIZE... | the_stack |
//#include <cstdio>
#include <stdio.h>
#include <vector>
#include <algorithm>
#include <math.h>
#include <omp.h>
#include <cuda.h>
#include <cuda_runtime_api.h>
using namespace std;
// texture<float, 2, cudaReadModeElementType> inTex;
texture<float4, 2, cudaReadModeElementType> inTex;
// texture<float, cudaTextureType... | the_stack |
#include <nvbio/basic/timer.h>
#include <nvbio/basic/shared_pointer.h>
#include <nvbio/io/sequence/sequence.h>
#include <nvbio/basic/dna.h>
#include <thrust/host_vector.h>
#include <thrust/device_vector.h>
#include <zlib/zlib.h>
#include <stdio.h>
#include <stdlib.h>
#include <vector>
#include <algorithm>
using namesp... | the_stack |
using namespace Yolo;
namespace nvinfer1
{
YoloLayerPlugin::YoloLayerPlugin(const int cudaThread /*= 512*/):mThreadCount(cudaThread)
{
mClassCount = CLASS_NUM;
mYoloKernel.clear();
mYoloKernel.push_back(yolo1);
mYoloKernel.push_back(yolo2);
mYoloKernel.push_back(yolo3);
... | the_stack |
#include "blockMatchingKernel.h"
#include "_reg_ReadWriteImage.h"
#include "_reg_tools.h"
#include <vector>
#include "_reg_maths.h"
//#define USE_TEST_KERNEL
////////////////////////////////////////////////////////////////////////////
////////////////////////////////////////////////////////////////////////////
/*
* ... | the_stack |
#include <nbla/array.hpp>
#include <nbla/cuda/common.hpp>
#include <nbla/cuda/function/pad.hpp>
#include <nbla/cuda/utils/atomic_add.cuh>
#include <nbla/variable.hpp>
namespace nbla {
using cuda::Index_t;
struct AxisParam {
Index_t x_stride;
Index_t y_stride;
Index_t y_shape;
struct {
Index_t first;
... | the_stack |
typedef uint32_t sph_u32;
#define C32(x) (x)
#define T32(x) (x)
#define INPUT_BLOCK_ADD do { \
B0 = T32(B0 + M0); \
B1 = T32(B1 + M1); \
B2 = T32(B2 + M2); \
B3 = T32(B3 + M3); \
B4 = T32(B4 + M4); \
B5 = T32(B5 + M5); \
B6 = T32(B6 + M6); \
B7 = T32(B7 + M7); \
B8 = T32(B8 + M8); \
B9 = T32(B9 + ... | the_stack |
#if !defined(ENABLE_GPU) | !defined(ENABLE_CUDNN)
#error "nnconv_cudnn.hpp cannot be compiled without GPU and CUDNN support."
#endif
#include "nnconv_cudnn.hpp"
#include "cudnnhelper.hpp"
#include "../datacu.hpp"
#include <assert.h>
#include <algorithm>
using namespace vl ;
#define CHECK(x) \
{ \
cudnnError = x ; \
... | the_stack |
#include <cudf/column/column.hpp>
#include <cudf/column/column_factories.hpp>
#include <cudf/detail/gather.hpp>
#include <cudf/detail/valid_if.cuh>
#include <cudf/filling.hpp>
#include <cudf/null_mask.hpp>
#include <cudf/scalar/scalar_factories.hpp>
#include <cudf/table/table.hpp>
#include <cudf/types.hpp>
#include <cu... | the_stack |
* Radix sorting problem instance
******************************************************************************/
#pragma once
#include "../util/spine.cuh"
#include "../util/basic_utils.cuh"
#include "../util/kernel_props.cuh"
#include "../util/error_utils.cuh"
#include "../util/cta_work_distribution.cuh"
#include ".... | the_stack |
#include "gpu_utils.cuh"
#include "nonbonded.hpp"
#include "vendored/hilbert.h"
#include "k_nonbonded.cuh"
#include <fstream>
#include <streambuf>
#include <string>
namespace timemachine {
template <typename RealType, bool Interpolated>
Nonbonded<RealType, Interpolated>::Nonbonded(
const std::vector<int> &exclu... | the_stack |
// #ifdef _HFS_CUDA_ON_
#include "../precomp.hpp"
#include "../slic/slic.hpp"
namespace cv { namespace hfs { namespace slic { namespace engines {
__global__ void cvtImgSpaceDevice(const Vector4u* inimg,
Vector2i img_size, Vector4f* outimg);
__global__ void initClusterCentersDevice(const Vector4f* inimg,
Ve... | the_stack |
using namespace std;
//==============================================================================
//
// Error handling helpers
//
//==============================================================================
static void stdDebugOutput(const string &msg)
{
cout << msg;
}
static NCVDebugO... | the_stack |
#include "gtest/gtest.h"
#include "keops_includes.h"
using namespace keops;
namespace {
TEST(tensordot, zeros){
__TYPE__ FA[4] = {4.4, 5.4, 6.2, 6.5};
__TYPE__ FB[4] = {1.4, 1.2, 1.5, 1.22};
auto x = Vi(0,4); // x is the second variable and represents a 3D vector, "i"-indexed.
auto y = Vj(1,4); // y i... | the_stack |
namespace cudatbx {
namespace scattering {
/* ==========================================================================
*/
cudatbx::scattering::direct_summation::direct_summation() {
// set host and device pointers to NULL
h_xyz = NULL;
h_solvent = NULL;
h_h = NULL;
h_rt = NULL;
h_weigh... | the_stack |
// Torch port:
// IMAGINE, Sergey Zagoruyko, Francisco Massa, 2015
#include "THC.h"
#include <algorithm>
#include <cfloat>
#include "common.h"
using std::max;
using std::min;
template <typename Dtype>
__global__ void ROIPoolForward(const int nthreads, const Dtype* bottom_data,
const Dtype spatial_scale, const... | the_stack |
#define XTALK_REGION_ONLY
/*
// xtalk calculation from excess hydrogen by neighbours
__global__ void SimpleXTalkNeighbourContributionAndAccumulation_LocalMem(// Here FL stands for flows
const unsigned short * RegionMask, //per Region
const unsigned short * bfMask, // per Bead
const unsigned short * bst... | the_stack |
#include <cusp/linear_operator.h>
#include <cusp/gallery/poisson.h>
#include <cusp/gallery/random.h>
#include <cusp/array2d.h>
#include <cusp/coo_matrix.h>
#include <cusp/csr_matrix.h>
#include <cusp/dia_matrix.h>
#include <cusp/ell_matrix.h>
#include <cusp/hyb_matrix.h>
#include <cusp/permutation_matrix.h>
#include ... | the_stack |
void blas_sgemm(cublasHandle_t handle,
const bool TransA, const bool TransB,
const int M, const int N, const int K,
const float alpha,
float *A, const int lda,
float *B, const int ldb,
const float beta,
... | the_stack |
#include "cuda_helper.h"
#undef SPH_ROTL32
#define SPH_ROTL32 ROTL32
static uint32_t *d_gnounce[MAX_GPUS];
static uint32_t *d_GNonce[MAX_GPUS];
__constant__ uint64_t pTarget[4];
#define shl(x, n) ((x) << (n))
#define shr(x, n) ((x) >> (n))
#define ss0(x) (shr((x), 1) ^ shl((x), 3) ^ SPH_ROTL32((x), 4) ^ SPH_ROTL3... | the_stack |
\brief Unit tests for thread-level GEMM
*/
#include "cutlass/arch/wmma.h"
#ifdef CUTLASS_ARCH_WMMA_SM70_ENABLED
#include "mma_pipelined_testbed.h"
#include "cutlass/gemm/threadblock/default_mma_core_wmma.h"
/// All tests use single staged (kStages=1) mma pipeline for the gemm mainloop
/// Test name format: SM[arch]_g... | the_stack |
#include <cugraph/graph_view.hpp>
#include <cugraph/matrix_partition_device_view.cuh>
#include <cugraph/partition_manager.hpp>
#include <cugraph/prims/row_col_properties.cuh>
#include <cugraph/utilities/dataframe_buffer.cuh>
#include <cugraph/utilities/device_comm.cuh>
#include <cugraph/utilities/error.hpp>
#include <c... | the_stack |
//#include <thrust/device_ptr.h>
//#include <thrust/scan.h>
__device__ __forceinline__ void decode_subsequence(
std::uint32_t subsequence_size,
std::uint32_t current_subsequence,
std::uint32_t subsequences_processed,
UNIT_TYPE mask,
std::uint32_t shift,
std::uint32_t start_bit,
std::uint32_... | the_stack |
#include <opencv2/cudafeatures2d.hpp>
#include "cuda_runtime.h"
#include "device_launch_parameters.h"
#include "labeling_algorithms.h"
#include "register.h"
#define BLOCK_X 8
#define BLOCK_Y 8
#define BLOCK_Z 4
using namespace cv;
namespace {
// Only use it with unsigned numeric types
template <typename T>
_... | the_stack |
////////////////////////////////////////////////////////////////////////////////////////////////////////////////
/// Depth bilateral filter
namespace kfusion {
namespace device {
__global__ void bilateral_kernel(const PtrStepSz<ushort> src, PtrStep<ushort> dst, const int ksz,
const flo... | the_stack |
* \file nms.cu
* \brief NMS Operator
* \author Yanghao Li
*/
#include <dmlc/logging.h>
#include <dmlc/parameter.h>
#include <mxnet/operator.h>
#include <mshadow/tensor.h>
#include <mshadow/cuda/reduce.cuh>
#include <thrust/sort.h>
#include <thrust/execution_policy.h>
#include <thrust/functional.h>
#include "../tensor... | the_stack |
#include <hash/unordered_multiset.cuh>
#include <rmm/cuda_stream_view.hpp>
#include <rmm/device_uvector.hpp>
#include <rmm/exec_policy.hpp>
#include <thrust/binary_search.h>
#include <thrust/fill.h>
#include <thrust/find.h>
#include <thrust/iterator/counting_iterator.h>
#include <thrust/pair.h>
#include <thrust/trans... | the_stack |
#include <cmath>
#include "caffe/common.hpp"
#include "caffe/util/math_functions.hpp"
#include "caffe/util/winograd.hpp"
namespace caffe {
template <>
void caffe_gpu_gemm<float>(const CBLAS_TRANSPOSE TransA,
const CBLAS_TRANSPOSE TransB, const int M, const int N, const int K,
const float alpha, const float* ... | the_stack |
#include <cuda.h>
#include <cuda_runtime.h>
namespace {
#define CUDA_1D_KERNEL_LOOP(i, n) \
for (size_t i = blockIdx.x * blockDim.x + threadIdx.x; i < (n); \
i += blockDim.x * gridDim.x)
// The number of cuda threads to use. 512 is used for backward compatibility
constexpr in... | the_stack |
template <int FS, int SB, int padding_l, typename scalar_t>
__global__ void lightconv_forward_kernel(
const scalar_t* input,
const scalar_t* filters,
int minibatch,
int sequenceLength,
int numFeatures,
int numFiltersInBlock,
scalar_t* output) {
const int tid = threadIdx.x;
const int batc... | the_stack |
#include <cuda.h>
#include <cuda_runtime.h>
#if CUDA_VERSION >= 9000 && __CUDA_ARCH__ >= 300
#undef __shfl
#define __shfl(var, srcLane, width) __shfl_sync(0xFFFFFFFFu, var, srcLane, width)
#endif
#include "cryptonight.h"
#define LONG_SHL32 19 // 1<<19 (uint32_t* index)
#define LONG_SHL64 18 // 1<<18 (uint64_t* index... | the_stack |
#include <assert.h>
#include <iostream>
#include <iomanip>
#include <chrono>
#include <vector>
#include <algorithm>
#include <stdlib.h>
#include <cuda.h>
using namespace std;
// to avoid integer overflow, n should not exceed this constant
constexpr int kMaxN = 31;
// kLimit = 100 million means we can use up to 3.2GB ... | the_stack |
#include "dragon/core/context_cuda.h"
#include "dragon/utils/device/common_cub.h"
#include "dragon/utils/math_functions.h"
#include "dragon/utils/op_kernels.h"
namespace dragon {
namespace kernels {
namespace {
#define LDG(x, i) __ldg(x + i)
#define LDG2(x, i) convert::To<AccT>(__ldg(x + i))
template <typename T, ... | the_stack |
#include <algorithm>
#include <cuda_runtime.h>
#include <device_launch_parameters.h>
#include <thrust/device_ptr.h>
#include <thrust/scan.h>
#include <thrust/sort.h>
#include <thrust/gather.h>
namespace cuba
{
namespace gpu
{
////////////////////////////////////////////////////////////////////////////////////
// Typ... | the_stack |
#define BLOCK_X 16
#define BLOCK_Y 16
#define PI 3.1415926535897932f
#define A 1103515245
#define C 12345
#define M INT_MAX
#define SCALE_FACTOR 300.0f
#ifndef BLOCK_SIZE
#define BLOCK_SIZE 256
#endif
#include "kernel_find_index.h"
#include "kernel_likelihood.h"
#include "kernel_normalize_weights.h"
#i... | the_stack |
extern "C" {
#include <stdint.h>
#include <memory.h>
}
#include <cuda_helper.h>
#include <cuda_vectors.h>
#define TPB52 1024
#define TPB50 384
#define NPT 2
#define NBN 2
static uint32_t *d_nonces[MAX_GPUS];
static uint32_t *h_nonces[MAX_GPUS];
__constant__ uint2 c_message48[6];
__constant__ uint2 c_mid[17];
__con... | the_stack |
namespace anakin {
namespace saber {
namespace {
template <ReduceType type>
class ReOp {
public:
__device__
static float compute(float a, float b) {
return -1.f;
}
};
template <>
__device__
float ReOp<Reduce_max>::compute(float a, float b) {
return ((a > b) ? a : b);
}
template <>
__device__
fl... | the_stack |
#define SEQ_LEN 33
__device__
inline double firstRepeatOffsetProb(const double probMult, const int maxRepeatOffset) {
if (probMult < 1 || probMult > 1)
return (1 - probMult) / (1 - pow(probMult, (double)maxRepeatOffset));
else
return 1.0 / maxRepeatOffset;
}
__device__
void maskProbableLetters(const int s... | the_stack |
#define _SIZE_T_DEFINED
#include <cuda.h>
#include <device_launch_parameters.h>
#include <texture_fetch_functions.h>
#include "float.h"
#include <builtin_types.h>
#include <vector_functions.h>
#include <math.h>
#include "../NeuralNetwork/Activation/ActivationFunction.cu"
extern "C"
{
__global__ void LSTMDeltaKern... | the_stack |
#include <cudf/utilities/traits.hpp>
#include <cmath>
namespace cudf {
namespace binops {
namespace compiled {
// All binary operations
namespace ops {
struct Add {
template <typename T1, typename T2>
CUDA_DEVICE_CALLABLE auto operator()(T1 const& lhs, T2 const& rhs) -> decltype(lhs + rhs)
{
return lhs + ... | the_stack |
#include <math.h> // required for fabs()
#include <float.h> // required for DBL_EPSILON
// Externally Defined Routines //
extern "C" __host__ __device__
double xChebyshev_Tn_Series(double x, const double a[], int degree);
// Int... | the_stack |
namespace MegBA {
namespace geo {
namespace {
template <typename T>
__global__ void AngleAxisToRotationKernel(
const int nItem, const int N, const T *valueDevicePtr0,
const T *valueDevicePtr1, const T *valueDevicePtr2, const T *gradDevicePtr0,
const T *gradDevicePtr1, const T *gradDevicePtr2, T *R0, T *R1, ... | the_stack |
#include "caffe2/operators/group_norm_op.h"
#include "caffe2/core/context_gpu.h"
#include "caffe2/utils/math.h"
#include "caffe2/utils/math/reduce.cuh"
namespace caffe2 {
namespace {
template <typename T>
__global__ void ComputeFusedParamsCUDAKernel(
const int N,
const int G,
const int K,
const T* m... | the_stack |
#include "cupoch/geometry/kdtree_flann.h"
#include "cupoch/geometry/pointcloud.h"
#include "cupoch/registration/feature.h"
#include "cupoch/utility/eigenvalue.h"
#include "cupoch/utility/console.h"
namespace cupoch {
namespace registration {
__constant__ float PST_RAD_45 = 0.78539816339744830961566084581988;
__consta... | the_stack |
* This sample demonstrates how use texture fetches in CUDA
*
* This sample takes an input PGM image (image_filename) and generates
* an output PGM image (image_filename_out). This CUDA kernel performs
* a simple 2D transform (rotation) on the texture coordinates (u,v).
*/
#include <shrUtils.h>
// includes, system
#i... | the_stack |
#include <cudaconv2.cuh>
/*
* Block size: 16x16.
* blockIdx.x determines case in batches of 16*imgsPerThread.
* blockIdx.y determines 4x4 image region in target image.
*
* threadIdx.x determines case.
* threadIdx.y determines pixel.
*
* hidActs: (numFilters, numModulesY, numModulesX, numImages)
* filters:... | the_stack |
#include "interpolation_cuda_kernel.cuh"
#include <ATen/ATen.h>
#include <ATen/NativeFunctions.h>
#include <ATen/Dispatch.h>
#include <ATen/cuda/CUDAApplyUtils.cuh>
#define min(a,b) ((a<b)?(a):(b))
#define max(a,b) ((a>b)?(a):(b))
#define DEBUG (0)
#ifndef BLOCKDIMX
#define BLOCKDIMX (32)
#endif
#ifndef BLOCKDIMY
... | the_stack |
// define to creat new layout on device instead of copying from host
//#define CREATE_POISSON_LUT_ON_DEVICE
#define DEBUG_TABLE_GAMMA 0
__device__ void
ComputeHydrogenForMultiFlowFit_dev(
int sId,
int flow_ndx,
int nucid,
float * nucRise,
float A,
float Krate,
float gain,
float SP,
float d,
int c... | the_stack |
using namespace onnxruntime::cuda;
namespace onnxruntime {
namespace contrib {
namespace cuda {
template <typename T>
__global__ void TransposeCtx(const int H, const bool reversed_bs, const T* input, T* output) {
// Input: BxNxSxH
// Output: BxSxNxH
int n = threadIdx.y;
int s = blockIdx.x;
int b = blockId... | the_stack |
#include "timer.h"
// CUDA libraries
#include <hip/hip_runtime.h>
/** Problem size along one side; total number of cells is this squared */
#define NUM 256
// block size
#define BLOCK_SIZE 128
#define Real float
#define ZERO 0.0f
#define ONE 1.0f
#define TWO 2.0f
/** SOR relaxation parameter */
const Real omega = ... | the_stack |
constexpr int NUM_PER_THREAD_REDUCE = 4;
constexpr int WARP_SIZE = 32;
constexpr int NUM_SHARED_SUM_INPUT = 6;
constexpr int NUM_SHARED_SUM_GAMMA = 3;
template <typename T>
inline __device__ T my_pow(T a, double b) {
return pow(a, static_cast<float>(b));
}
template <>
inline __device__ half my_pow(half a, double b)... | the_stack |
#include "LabelIslandSortArea.h"
#include <iostream>
using namespace std;
// 宏:MAX_PAIRS_NUM
//(面积值-标记值)键值对的个数。
#ifndef MAX_PAIRS_NUM
#define MAX_PAIRS_NUM 256
#endif
// 宏:SORT_ARRAY_TYPE_ASC
// 排序标识,升序排序。
#ifndef SORT_ARRAY_TYPE_ASC
#define SORT_ARRAY_TYPE_ASC 2
#endif
// 宏:SORT_ARRAY_TYPE_DESC
// 排序标识,降序排序。
... | the_stack |
#include <GL/glew.h>
#if defined(__APPLE__) || defined(MACOSX)
#include <GLUT/glut.h>
#else
#include <GL/freeglut.h>
#endif
// CUDA standard includes
#include <cuda_runtime.h>
#include <cuda_gl_interop.h>
#include "bodysystem.h"
__constant__ float softeningSquared;
__constant__ double softeningSquared_fp64;
cudaErr... | the_stack |
#ifndef _BLURKERNEL_H_
#define _BLURKERNEL_H_
#include "gpu/image/blur.hpp"
#include "backend/common/imageOps.hpp"
#include "backend/common/image/blurdef.h"
namespace VideoStitch {
namespace Image {
/**
* Accumulator values for several types.
*/
template <typename U>
struct AccumT {};
/**
* Accumulator values f... | the_stack |
#include <ATen/div_rtn.h>
#include <ATen/AccumulateType.h>
#include <ATen/cuda/CUDABlas.h>
#include <ATen/native/ConvUtils.h>
#include <ATen/native/Resize.h>
#include <ATen/native/cuda/im2col.cuh>
namespace at { namespace native {
namespace {
void slow_conv2d_shape_check(
const Tensor& input, const Tensor& grad_o... | the_stack |
#include <thrust/copy.h>
#include <thrust/execution_policy.h>
#include <thrust/host_vector.h>
#include <thrust/sort.h>
#define FORCE_CPU_VERSION
#include "BufferEntryUtils.h"
#undef FORCE_CPU_VERSION
namespace {
template <class K, class V, class I>
std::vector<uint32_t> do_radix_sort(const ExecutorDeviceType device_... | the_stack |
namespace tl = treelite;
/* All functions and classes defined in this anonymous namespace are strictly
* for internal use by GPUTreeSHAP. */
namespace {
template <typename ThresholdType>
struct SplitCondition {
SplitCondition() = default;
SplitCondition(ThresholdType feature_lower_bound,
Thresho... | the_stack |
using namespace facebook::cuda;
namespace facebook { namespace deeplearning { namespace torch {
namespace detail {
extern __shared__ float pShared[];
template <int BatchSize, typename T>
__launch_bounds__(256, 6)
__global__ void updateOutputBatch(DeviceTensor<T, 4> input,
DeviceTens... | the_stack |
#include <nbla/array.hpp>
#include <nbla/cuda/common.hpp>
#include <nbla/cuda/function/max_pooling_backward.hpp>
#include <nbla/cuda/utils/atomic_add.cuh>
#include <nbla/cuda/utils/nd_index.cuh>
#include <nbla/variable.hpp>
#include <thrust/device_ptr.h>
#include <thrust/device_vector.h>
namespace nbla {
namespace m... | the_stack |
#include <cuda.h>
#include <cuda_runtime.h>
#include <math.h>
#include <float.h>
#include <vector>
#define CUDA_1D_KERNEL_LOOP(i, n) \
for (int i = blockIdx.x * blockDim.x + threadIdx.x; i < n; \
i += blockDim.x * gridDim.x)
// CUDA: grid stride looping
#define CUDA_KERNEL_LO... | the_stack |
using namespace std;
#define DEBUG_SYNC 1 //forces sync after each kernel/async call to guarantee correct error catching
#define DEBUG_OUTPUT 0 //fully verbose
#define DEBUG_REGION 999999 // region to print for debugging, if DEBUG_REGION_ALL is set this value is ignored
#define DEBUG_REGION_ALL 0 // if set to 1 al... | the_stack |
#pragma once
#include <gunrock/oprtr/compacted_cull_filter/kernel_policy.cuh>
#include <gunrock/oprtr/cull_filter/cta.cuh>
namespace gunrock {
namespace oprtr {
namespace compacted_cull_filter {
template <typename KernelPolicy, typename Problem, typename Functor>
struct ThreadWork {
typedef typename Problem::Verte... | the_stack |
#pragma once
#include <gunrock/util/srts_grid.cuh>
#include <gunrock/util/reduction/soa/cooperative_soa_reduction.cuh>
#include <gunrock/util/scan/soa/serial_soa_scan.cuh>
#include <gunrock/util/scan/soa/warp_soa_scan.cuh>
namespace gunrock {
namespace util {
namespace scan {
namespace soa {
/**
* Cooperative SOA r... | the_stack |
@page
@model Sweetalert2Model
@{
ViewData["Title"] = "SweetAlert2";
ViewData["PageName"] = "notifications_sweetalert2";
ViewData["Category1"] = "Notifications";
ViewData["Heading"] = "<i class='subheader-icon fal fa-exclamation-circle'></i> SweetAlert2 <sup class='badge badge-primary fw-500'>ADDON</sup>";
ViewDat... | the_stack |
@using lsc.Common
@using lsc.Model
@using lsc.Model.Enume
@model EnterCustomer
@{
ViewData["Title"] = "添加客户";
Layout = "~/Pages/_Layout.cshtml";
List<DistrictInfo> ProvinceList = ViewBag.ProvinceList;
}
<blockquote class="layui-elem-quote">
添加客户信息
</blockquote>
<div class="manage-form-container">
... | the_stack |
@{
ViewBag.Title = "User";
}
<div style="width:100%;height:100%">
<div class="panel-body" style="padding-bottom:0px;">
<div class="panel panel-default">
<div class="panel-heading">查询条件</div>
<div class="panel-body">
<form id="formSearch" class="form-horizontal">... | the_stack |
@model SmartAdmin.Domain.Models.Notification
@{
/**/
ViewBag.Title = "消息推送";
ViewData["PageName"] = "notifications_index";
ViewData["Heading"] = "<i class='fal fa-comments-alt text-primary'></i> 消息推送";
ViewData["Category1"] = "系统管理";
ViewData["PageDescription"] = "";
}
@section HeadBlock {
<link href=... | the_stack |
@using lsc.Model
@using lsc.Model.Enume
@model List<lsc.Model.EnterCustomer>
@{
ViewData["Title"] = "客户管理";
Layout = "~/Pages/_Layout.cshtml";
List<DistrictInfo> ProvinceList = ViewBag.ProvinceList;
List<UserRoleJurisdiction> userrolejurlist = ViewData["userrolejurlist"] as List<UserRoleJurisdiction>;... | the_stack |
@page
@model AlteditorModel
@{
ViewData["Title"] = "AltEditor (beta)";
ViewData["PageName"] = "datatables_alteditor";
ViewData["Category1"] = "Datatables";
ViewData["Heading"] = "<i class='subheader-icon fal fa-table'></i> DataTables: <span class='fw-300'>AltEditor (beta)</span> <sup class='badge badge-primary fw-... | the_stack |
#addin nuget:?package=Cake.Coveralls&version=1.0.1
#addin nuget:?package=Cake.FileHelpers&version=4.0.1
//#addin nuget:?package=Cake.Issues&version=0.9.1
#addin nuget:?package=Cake.AppVeyor&version=5.0.1
#addin nuget:?package=Cake.ReSharperReports&version=0.11.1
// TOOLS
#tool nuget:?package=GitReleaseManager&version=... | the_stack |
#tool nuget:?package=vswhere&version=2.7.1
// Cake Addins
#addin nuget:?package=Cake.FileHelpers&version=3.2.1
#addin nuget:?package=Newtonsoft.Json&version=11.0.2
using System;
using System.Text.RegularExpressions;
using System.Xml;
using System.Xml.Linq;
using Newtonsoft.Json;
using Newtonsoft.Json.Linq;
var TARG... | the_stack |
@model puck.core.Models.PuckImage
@using puck.core.Models.EditorSettings
@{
var settings = this.PuckEditorSettings<CropsEditorSettings>(inherit:false,modelTypeOverride:typeof(puck.core.Base.BaseModel))
?? new CropsEditorSettings()
{
Crops=new List<puck.core.Models.CropInfo>()
};... | the_stack |
@page
@model SearchModel
@{
ViewData["Title"] = "Search";
ViewData["PageName"] = "page_search";
ViewData["Category1"] = "Page Views";
ViewData["PreemptiveClass"] = "layout-composed";
}
@section HeadBlock {
<link rel="stylesheet" media="screen, print" href="~/css/fa-solid.css">
<link rel="stylesheet" media="screen,... | the_stack |
@{
ViewBag.MainMenu = "Settings";
ViewBag.LeftMenu = "Material";
ViewBag.Title = "图文消息";
Layout = "~/Views/Shared/_Layout.cshtml";
}
<link rel="stylesheet" type="text/css" href="/Scripts/wangEditor/css/wangEditor.min.css">
<script type="text/javascript" src='/Scripts/wangEditor/js/wangEditor.min.js'>... | the_stack |
// https://www.nuget.org/api/v2/package/Newtonsoft.Json/12.0.2
#r "sha256:b9b4e633ea6c728bad5f7cbbef7f8b842f7e10181731dbe5ec3cd995a6f60287"
// from elm-fullstack-separate-assemblies-4505d5fa0951dbb5d83383b17058704c58ebc674-linux-x64.zip
#r "sha256:67d1550a5b06e9b361fdc9220062dd960e036e7daaa063e92380d186f93089cf"
usi... | the_stack |
@model Sheng.WeixinConstruction.Client.Shell.Models.LuckyTicketDrawResultViewModel
@{
ViewBag.SubTitle = "活动";
Layout = "~/Views/Shared/_LayoutBlank.cshtml";
}
<style type="text/css">
body {
margin-bottom: 0.55rem;
}
.campaignName {
font-size: 0.14rem;
font-weight: bold;
... | the_stack |
@using puck.core.Abstract.EditorSettings
@using puck.core.Models.EditorSettings
@using puck.core.Models.EditorSettings.Attributes
@model List<puck.core.Models.PuckReference>
@{
I_Content_Picker_Settings settings = this.PuckEditorSettings<ContentPickerEditorSettingsAttribute>();
if (settings == null)
{
... | the_stack |
using System;
using System.Diagnostics;
using System.IO;
using System.Xml;
using System.Collections.Generic;
using System.Reflection;
using System.Runtime.InteropServices;
using System.Linq;
using Foundation;
using AppKit;
using ObjCRuntime;
// -------------------------------------------------------------------------... | the_stack |
@{
ViewBag.Title = "InformationItemEdit";
Layout = "~/Views/Shared/_LayoutBlank.cshtml";
}
<link rel="stylesheet" type="text/css" href="/Scripts/wangEditor/css/wangEditor.min.css">
<script type="text/javascript" src='/Scripts/wangEditor/js/wangEditor.min.js'></script>
<style>
.wangEditor-container .wangE... | the_stack |
@using System.Globalization
@using Abp.Configuration
@using Abp.Configuration.Startup
@using Abp.Web.Security.AntiForgery
@using Volo.PostgreSqlDemo
@using Volo.PostgreSqlDemo.Configuration
@using Volo.PostgreSqlDemo.SignalR
@using Volo.PostgreSqlDemo.Web.Resources
@using Volo.PostgreSqlDemo.Web.Views.Shared.Component... | the_stack |
@model MOE.Common.Models.ViewModel.LinkPivotResultViewModel
<table id="AdjustmentTable" class="table table-striped table-bordered table-condensed">
<caption>Adjustments</caption>
<thead>
<tr>
<th class="vertical-center text-center" scope="col">Link</th>
<th class="vertical-cent... | the_stack |
@{
ViewBag.Title = "API Documentation";
var siteBaseUrl = "https://openchargemap.org/site/";
var apiBaseUrl = "https://api.openchargemap.io/v3/";
var apiSandboxBaseUrl = "https://sandbox.api.openchargemap.io/v2/";
}
<link rel="stylesheet" href="~/Content/docs.css" type="text/css" />
<div class="row"... | the_stack |
@model LoginViewModel
@inject SignInManager<UserEntity> SignInManager
@{
Layout = "~/Views/Shared/_Root.cshtml";
ViewData["Title"] = Localizer["Log in"];
}
<div id="content">
<div id="logo">
<svg version="1.1" id="layer_1" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlin... | the_stack |
@model SmartAdmin.WebUI.Data.Models.Tenant
@{
/**/
ViewData["Title"] = "租户管理";
ViewData["PageName"] = "tenants_index";
ViewData["Heading"] = "<i class='fal fa-users text-primary'></i> 租户管理";
ViewData["Category1"] = "系统管理";
ViewData["PageDescription"] = "";
}
@section HeadBlock {
<link href="~/js/easyui/... | the_stack |
@using System.Activities.Statements
<!DOCTYPE html>
<html >
<head>
<meta charset="utf-8" />
<meta name="viewport" content="width=device-width" />
<meta http-equiv="X-UA-Compatible" content="IE=edge">
<title>@ViewBag.Title</title>
<meta content="餐饮软件,点餐宝,点餐系统" name="keywords" />
<meta content... | the_stack |
@{
Layout = null;
}
@model MIMS.Entity.Model.PSS_PurchasePlan
<!DOCTYPE html>
<html>
<head>
<meta name="viewport" content="width=device-width" />
<title>PurchasePlanDetail</title>
@*Easyui需要引入的文件*@
<script type="text/javascript" src="~/Content/jquery-easyui-1.4.5/jquery.min.js"></script>
<scr... | the_stack |
@page
@model GeneralModel
@{
ViewData["Title"] = "General Docs";
ViewData["PageName"] = "docs_general";
ViewData["Heading"] = "<i class='subheader-icon fal fa-book'></i> Documentation: <span class='fw-300'>General Docs</span>";
ViewData["Category1"] = "Documentation";
ViewData["PageDescription"] = "Product docume... | the_stack |
@model Xms.Web.Customize.Models.SolutionComponentDialogModel
@{
Layout = null;
}
<!-- (Modal) -->
<div class="modal fade" id="solutionComponentModal" tabindex="-1" role="dialog"
aria-labelledby="solutionComponentModalLabel" aria-hidden="true">
<link href="~/content/js/jquery-ui-1.10.3/themes/base/jquery... | the_stack |
@using NewLife.Common;
@using NewLife.Cube.WebMiddleware;
@using NewLife.Cube.Charts;
@{
var set = NewLife.Cube.Setting.Current;
var set2 = ViewBag.PageSetting as PageSetting;
var title = ViewBag.Title + "";
if (title != "" && !title.Contains(" - "))
{
title += " - " + NewLife.Common.SysCo... | the_stack |
@page
@model C3Model
@{
ViewData["Title"] = "C3 Charts";
ViewData["PageName"] = "statistics_c3";
ViewData["Category1"] = "Statistics";
ViewData["Heading"] = "<i class='subheader-icon fal fa-chart-pie'></i> C3 Charts<sup class='badge badge-primary fw-500'>ADDON</sup>";
ViewData["PageDescription"] = "Dygraphs is a ... | the_stack |
@using lsc.Common
@model lsc.Model.EnterCustContacts
@{
ViewData["Title"] = "客户联系人";
Layout = "~/Pages/_LayoutNone.cshtml";
int type = ViewBag.type;
if (type!=0)
{
Layout = "~/Pages/_Layout.cshtml";
}
//if (Model!=null)
//{
// Layout = "~/Pages/_LayoutNone.cshtml";
... | the_stack |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.