hip
stringlengths
140
3.32k
cuda
stringlengths
84
3.33k
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/common_gpu.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/tensor_protos_db_input.h" namespace caffe2 { REGISTER_HIP_OPERATOR(TensorProtosDBInput, TensorProtosDBInput<HIPContext>); } // namespace caffe2 ###
#include "caffe2/core/common_gpu.h" #include "caffe2/core/context_gpu.h" #include "caffe2/operators/tensor_protos_db_input.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(TensorProtosDBInput, TensorProtosDBInput<CUDAContext>); } // namespace caffe2 ###
// !!! This is a file automatically generated by hipify!!! #include "hip/hip_runtime.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/thresholded_relu_op.h" namespace caffe2 { namespace { template <typename T> __global__ void ThresholdedReluKernel(const int N, const T* X, T* Y, T alpha_) { HIP...
#include "caffe2/core/context_gpu.h" #include "caffe2/operators/thresholded_relu_op.h" namespace caffe2 { namespace { template <typename T> __global__ void ThresholdedReluKernel(const int N, const T* X, T* Y, T alpha_) { CUDA_1D_KERNEL_LOOP(i, N) { Y[i] = X[i] > alpha_ ? X[i] : 0; } } template <typename T> _...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/operators/transpose_op.h" #include "caffe2/core/hip/context_gpu.h" namespace caffe2 { REGISTER_HIP_OPERATOR(Transpose, TransposeOp<HIPContext>); } // namespace caffe2 ###
#include "caffe2/operators/transpose_op.h" #include "caffe2/core/context_gpu.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(Transpose, TransposeOp<CUDAContext>); } // namespace caffe2 ###
#include "hip/hip_runtime.h" #include "caffe2/operators/unique_ops.h" #include <thrust/device_vector.h> #include <thrust/sequence.h> #include <thrust/sort.h> #include <thrust/system/hip/execution_policy.h> #include <thrust/unique.h> #include <thrust/version.h> #include "caffe2/core/hip/context_gpu.h" namespace caffe...
#include "caffe2/operators/unique_ops.h" #include <thrust/device_vector.h> #include <thrust/sequence.h> #include <thrust/sort.h> #include <thrust/system/cuda/execution_policy.h> #include <thrust/unique.h> #include <thrust/version.h> #include "caffe2/core/context_gpu.h" namespace caffe2 { #if THRUST_VERSION >= 100800 ...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/operators/unsafe_coalesce.h" #include "caffe2/core/hip/context_gpu.h" namespace caffe2 { REGISTER_HIP_OPERATOR(UnsafeCoalesce, UnsafeCoalesceOp<HIPContext>); } ###
#include "caffe2/operators/unsafe_coalesce.h" #include "caffe2/core/context_gpu.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(UnsafeCoalesce, UnsafeCoalesceOp<CUDAContext>); } ###
// !!! This is a file automatically generated by hipify!!! #include <iostream> #include "caffe2/core/context.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/core/flags.h" #include "caffe2/operators/utility_ops.h" #include <gtest/gtest.h> C10_DECLARE_string(caffe_test_root); namespace caffe2 { static v...
#include <iostream> #include "caffe2/core/context.h" #include "caffe2/core/context_gpu.h" #include "caffe2/core/flags.h" #include "caffe2/operators/utility_ops.h" #include <gtest/gtest.h> C10_DECLARE_string(caffe_test_root); namespace caffe2 { static void AddConstInput( const vector<int64_t>& shape, const ...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/operators/while_op.h" #include "caffe2/core/hip/context_gpu.h" namespace caffe2 { REGISTER_HIP_OPERATOR(While, WhileOp<HIPContext>); } // namespace caffe2 ###
#include "caffe2/operators/while_op.h" #include "caffe2/core/context_gpu.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(While, WhileOp<CUDAContext>); } // namespace caffe2 ###
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/zero_gradient_op.h" namespace caffe2 { REGISTER_HIP_OPERATOR(ZeroGradient, ZeroGradientOp<HIPContext>); } ###
#include "caffe2/core/context_gpu.h" #include "caffe2/operators/zero_gradient_op.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(ZeroGradient, ZeroGradientOp<CUDAContext>); } ###
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_blob_fetcher_op.h" namespace caffe2 { REGISTER_HIP_OPERATOR( RecurrentNetworkBlobFetcher, RecurrentNetworkBlobFetcherOp<HIPContext>); } // namespace caffe2 ###
#include "caffe2/core/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_blob_fetcher_op.h" namespace caffe2 { REGISTER_CUDA_OPERATOR( RecurrentNetworkBlobFetcher, RecurrentNetworkBlobFetcherOp<CUDAContext>); } // namespace caffe2 ###
// !!! This is a file automatically generated by hipify!!! #ifndef CAFFE2_OPERATORS_RECURRENT_NETWORK_GPU_EXECUTOR_H_ #define CAFFE2_OPERATORS_RECURRENT_NETWORK_GPU_EXECUTOR_H_ #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_executor.h" #include <map> namespace caffe2 { c...
#ifndef CAFFE2_OPERATORS_RECURRENT_NETWORK_GPU_EXECUTOR_H_ #define CAFFE2_OPERATORS_RECURRENT_NETWORK_GPU_EXECUTOR_H_ #include "caffe2/core/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_executor.h" #include <map> namespace caffe2 { class CUDARecurrentNetworkExecutor : public RecurrentNetworkExec...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #define _USE_MATH_DEFINES #include <ATen/native/Activation.h> #include <cmath> #include <thrust/tuple.h> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/core/TensorBase.h> #include <c10/core/Scal...
#define TORCH_ASSERT_NO_OPERATORS #define _USE_MATH_DEFINES #include <ATen/native/Activation.h> #include <cmath> #include <thrust/tuple.h> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/core/TensorBase.h> #include <c10/core/Scalar.h> #include <c10/cuda/CUDAMathCompat.h> #include <ATen/c...
// !!! This is a file automatically generated by hipify!!! #include "hip/hip_runtime.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_op.h" namespace caffe2 { namespace detail { template <typename T, typename Context> void initializeRecurrentInput( const RecurrentInpu...
#include "caffe2/core/context_gpu.h" #include "caffe2/operators/rnn/recurrent_network_op.h" namespace caffe2 { namespace detail { template <typename T, typename Context> void initializeRecurrentInput( const RecurrentInput& rc, int32_t seqLen, int32_t batchSize, Workspace* ws, Context* context); ...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/queue/queue_ops.h" #include "caffe2/utils/math.h" #include "caffe2/core/hip/context_gpu.h" namespace caffe2 { REGISTER_HIP_OPERATOR(CreateBlobsQueue, CreateBlobsQueueOp<HIPContext>); REGISTER_HIP_OPERATOR(EnqueueBlobs, EnqueueBlobsOp<HIPCon...
#include "caffe2/queue/queue_ops.h" #include "caffe2/utils/math.h" #include "caffe2/core/context_gpu.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(CreateBlobsQueue, CreateBlobsQueueOp<CUDAContext>); REGISTER_CUDA_OPERATOR(EnqueueBlobs, EnqueueBlobsOp<CUDAContext>); REGISTER_CUDA_OPERATOR(DequeueBlobs, DequeueBlobsOp...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/context_gpu.h" #include "caffe2/sgd/iter_op.h" namespace caffe2 { REGISTER_HIP_OPERATOR(Iter, IterOp<HIPContext>); REGISTER_HIP_OPERATOR(AtomicIter, AtomicIterOp<HIPContext>); } // namespace caffe2 ###
#include "caffe2/core/context_gpu.h" #include "caffe2/sgd/iter_op.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(Iter, IterOp<CUDAContext>); REGISTER_CUDA_OPERATOR(AtomicIter, AtomicIterOp<CUDAContext>); } // namespace caffe2 ###
// !!! This is a file automatically generated by hipify!!! #include "hip/hip_runtime.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/sgd/lars_op.h" namespace caffe2 { __global__ void ComputeLearningRateKernel( const float* wd, const float* trust, const float* lr_max, float offset, flo...
#include "caffe2/core/context_gpu.h" #include "caffe2/sgd/lars_op.h" namespace caffe2 { __global__ void ComputeLearningRateKernel( const float* wd, const float* trust, const float* lr_max, float offset, float lr_min, float* X_norm, float* dX_norm, float* lr_rescaled) { float val = 1....
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/context_gpu.h" #include "caffe2/sgd/learning_rate_op.h" namespace caffe2 { REGISTER_HIP_OPERATOR(LearningRate, LearningRateOp<float, HIPContext>); } // namespace caffe2 ###
#include "caffe2/core/context_gpu.h" #include "caffe2/sgd/learning_rate_op.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(LearningRate, LearningRateOp<float, CUDAContext>); } // namespace caffe2 ###
// !!! This is a file automatically generated by hipify!!! #include "hip/hip_runtime.h" #include "caffe2/sgd/rmsprop_op.h" #include "caffe2/core/hip/common_gpu.h" #include "caffe2/core/hip/context_gpu.h" namespace caffe2 { __global__ void RmsPropUpdate( int N, const float* g, const float* ms, const f...
#include "caffe2/sgd/rmsprop_op.h" #include "caffe2/core/common_gpu.h" #include "caffe2/core/context_gpu.h" namespace caffe2 { __global__ void RmsPropUpdate( int N, const float* g, const float* ms, const float* mom, float* ng, float* nms, float* nmom, float decay, float momentum, ...
// !!! This is a file automatically generated by hipify!!! #include "caffe2/core/hip/common_gpu.h" #include "caffe2/core/hip/context_gpu.h" #include "caffe2/sgd/weight_scale_op.h" namespace caffe2 { REGISTER_HIP_OPERATOR(WeightScale, WeightScaleOp<HIPContext>); template <typename T> void weight_scale_update_kernel( ...
#include "caffe2/core/common_gpu.h" #include "caffe2/core/context_gpu.h" #include "caffe2/sgd/weight_scale_op.h" namespace caffe2 { REGISTER_CUDA_OPERATOR(WeightScale, WeightScaleOp<CUDAContext>); template <typename T> void weight_scale_update_kernel( int N, const T* w, const T& scale, int64_t iter, ...
// !!! This is a file automatically generated by hipify!!! #pragma once // cub sort support for CUB_WRAPPED_NAMESPACE is added to cub 1.13.1 in: // https://github.com/NVIDIA/cub/pull/326 // CUB_WRAPPED_NAMESPACE is defined globally in cmake/Dependencies.cmake // starting from HIP 11.5 #if defined(CUB_WRAPPED_NAMESPAC...
#pragma once // cub sort support for CUB_WRAPPED_NAMESPACE is added to cub 1.13.1 in: // https://github.com/NVIDIA/cub/pull/326 // CUB_WRAPPED_NAMESPACE is defined globally in cmake/Dependencies.cmake // starting from CUDA 11.5 #if defined(CUB_WRAPPED_NAMESPACE) || defined(THRUST_CUB_WRAPPED_NAMESPACE) #define USE_GL...
// !!! This is a file automatically generated by hipify!!! #ifndef CAFFE2_UTILS_GPU_ATOMICS_H_ #define CAFFE2_UTILS_GPU_ATOMICS_H_ #include <hip/hip_runtime.h> namespace caffe2 { namespace { template <typename T> inline __device__ void gpu_atomic_add(T* address, const T val) { atomicAdd(address, val); } templat...
#ifndef CAFFE2_UTILS_GPU_ATOMICS_H_ #define CAFFE2_UTILS_GPU_ATOMICS_H_ #include <cuda_runtime.h> namespace caffe2 { namespace { template <typename T> inline __device__ void gpu_atomic_add(T* address, const T val) { atomicAdd(address, val); } template <> inline __device__ void gpu_atomic_add(float* address, con...
#include "hip/hip_runtime.h" #ifndef CAFFE2_UTILS_GPU_SCAN_UTILS_H_ #define CAFFE2_UTILS_GPU_SCAN_UTILS_H_ #include "caffe2/utils/hip/GpuDefs.cuh" namespace caffe2 { template <typename T, bool KillWARDependency, class BinaryFunction> __device__ void inclusivePrefixScan(T* smem, T in, T* out, BinaryFunction binop)...
#ifndef CAFFE2_UTILS_GPU_SCAN_UTILS_H_ #define CAFFE2_UTILS_GPU_SCAN_UTILS_H_ #include "caffe2/utils/GpuDefs.cuh" namespace caffe2 { template <typename T, bool KillWARDependency, class BinaryFunction> __device__ void inclusivePrefixScan(T* smem, T in, T* out, BinaryFunction binop) { smem[threadIdx.x] = in; _...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPConfig.h> #include <ATen/hip\cub.cuh> namespace at { namespace hip { namespace cub { namespace detail { template <typename key_t, int value_size> void radix_sort_pairs_impl( const key_t* keys_in, key_t* keys_out, const OpaqueType<value_size>* values_in, Opaque...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAConfig.h> #include <ATen/cuda/cub.cuh> namespace at { namespace cuda { namespace cub { namespace detail { template <typename key_t, int value_size> void radix_sort_pairs_impl( const key_t* keys_in, key_t* keys_out, const OpaqueType<value_size>* values_in, Opa...
#include "hip/hip_runtime.h" #ifndef CAFFE2_UTILS_MATH_REDUCE_CUH_ #define CAFFE2_UTILS_MATH_REDUCE_CUH_ #include "caffe2/utils/cub_namespace.cuh" #include <hipcub/hipcub.hpp> #include "caffe2/core/hip/common_gpu.h" namespace caffe2 { template <typename T> using BlockReduce = hipcub::BlockReduce<T, CAFFE_HIP_NUM_THRE...
#ifndef CAFFE2_UTILS_MATH_REDUCE_CUH_ #define CAFFE2_UTILS_MATH_REDUCE_CUH_ #include "caffe2/utils/cub_namespace.cuh" #include <cub/block/block_reduce.cuh> #include "caffe2/core/common_gpu.h" namespace caffe2 { template <typename T> using BlockReduce = cub::BlockReduce<T, CAFFE_CUDA_NUM_THREADS>; template <typename T,...
// !!! This is a file automatically generated by hipify!!! #include <caffe2/core/hip/common_gpu.h> #include <caffe2/core/hip/context_gpu.h> #include <caffe2/video/video_input_op.h> namespace caffe2 { REGISTER_HIP_OPERATOR(VideoInput, VideoInputOp<HIPContext>); } // namespace caffe2 ###
#include <caffe2/core/common_gpu.h> #include <caffe2/core/context_gpu.h> #include <caffe2/video/video_input_op.h> namespace caffe2 { REGISTER_CUDA_OPERATOR(VideoInput, VideoInputOp<CUDAContext>); } // namespace caffe2 ###
#include <cfloat> #include "caffe2/core/hip/context_gpu.h" #include "modules/detectron/sample_as_op.h" #include <stdio.h> namespace caffe2 { template <> bool SampleAsOp<float, HIPContext>::RunOnDevice() { auto& X = Input(0); auto& L = Input(1); CAFFE_ENFORCE( X.dim32(0) == L.dim32(0), "X.dim32(0) must be eq...
#include <cfloat> #include "caffe2/core/context_gpu.h" #include "modules/detectron/sample_as_op.h" #include <stdio.h> namespace caffe2 { template <> bool SampleAsOp<float, CUDAContext>::RunOnDevice() { auto& X = Input(0); auto& L = Input(1); CAFFE_ENFORCE( X.dim32(0) == L.dim32(0), "X.dim32(0) must be equal ...
// !!! This is a file automatically generated by hipify!!! // Copyright (c) Meta Platforms, Inc. and affiliates. // // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. #pragma once #include <ATen/ATen.h> #include <vector> namespace torc...
// Copyright (c) Meta Platforms, Inc. and affiliates. // // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. #pragma once #include <ATen/ATen.h> #include <vector> namespace torch { namespace distributed { namespace c10d { namespace quan...
// !!! This is a file automatically generated by hipify!!! #define __NVFUSER_BFLOAT_TO_US(var) *(reinterpret_cast<unsigned short*>(&(var))) #define __NVFUSER_BFLOAT_TO_CUS(var) \ *(reinterpret_cast<const unsigned short*>(&(var))) struct __bfloat; __device__ __bfloat __float2bfloat(const float); struct __align__(2...
#define __NVFUSER_BFLOAT_TO_US(var) *(reinterpret_cast<unsigned short*>(&(var))) #define __NVFUSER_BFLOAT_TO_CUS(var) \ *(reinterpret_cast<const unsigned short*>(&(var))) struct __bfloat; __device__ __bfloat __float2bfloat(const float); struct __align__(2) __bfloat { __bfloat() = default; __device__ __bfloat...
// !!! This is a file automatically generated by hipify!!! struct __align__(2) __bfloat { __bfloat() = default; inline __device__ __bfloat(const float f) { if (f != f) { __x = uint16_t(0x7FC0); } else { union { uint32_t U32; float F32; }; F32 = f; uint32_t r...
struct __align__(2) __bfloat { __bfloat() = default; inline __device__ __bfloat(const float f) { if (f != f) { __x = uint16_t(0x7FC0); } else { union { uint32_t U32; float F32; }; F32 = f; uint32_t rounding_bias = ((U32 >> 16) & 1) + uint32_t(0x7FFF); ...
// !!! This is a file automatically generated by hipify!!! // Default block synchronization. Just use __barrier_sync namespace block_sync { __forceinline__ __device__ void init() {} // Thread-block synchronization __forceinline__ __device__ void sync() { __barrier_sync(0); } } // namespace block_sync ###
// Default block synchronization. Just use __barrier_sync namespace block_sync { __forceinline__ __device__ void init() {} // Thread-block synchronization __forceinline__ __device__ void sync() { __barrier_sync(0); } } // namespace block_sync ###
// !!! This is a file automatically generated by hipify!!! #define __NVFUSER_HALF_TO_US(var) *(reinterpret_cast<unsigned short*>(&(var))) #define __NVFUSER_HALF_TO_CUS(var) \ *(reinterpret_cast<const unsigned short*>(&(var))) struct __half; __device__ __half __float2half(const float); struct __align__(2) __half {...
#define __NVFUSER_HALF_TO_US(var) *(reinterpret_cast<unsigned short*>(&(var))) #define __NVFUSER_HALF_TO_CUS(var) \ *(reinterpret_cast<const unsigned short*>(&(var))) struct __half; __device__ __half __float2half(const float); struct __align__(2) __half { __half() = default; __device__ __half(const float f) ...
namespace fused_reduction { template < int NumVals, typename DataTypeT, typename IndexTypeT, template <int, typename> typename MakeTuple> struct WelfordTripletTuple { static constexpr int num_vals = NumVals; using DataType = DataTypeT; using IndexType = IndexTypeT; using DataTuple = typename MakeTuple<...
namespace fused_reduction { template < int NumVals, typename DataTypeT, typename IndexTypeT, template <int, typename> typename MakeTuple> struct WelfordTripletTuple { static constexpr int num_vals = NumVals; using DataType = DataTypeT; using IndexType = IndexTypeT; using DataTuple = typename MakeTuple<N...
// !!! This is a file automatically generated by hipify!!! namespace grid_broadcast { // Broadcasts per-thread values across threads and blocks. // // Function parameters: // - out: Per-thread output location // - inp_val: Per-thread input value // - work_buf: Temporary buffer for communication across threads/blocks ...
namespace grid_broadcast { // Broadcasts per-thread values across threads and blocks. // // Function parameters: // - out: Per-thread output location // - inp_val: Per-thread input value // - work_buf: Temporary buffer for communication across threads/blocks // - sync_flags: A vector of integers for synchronizations ...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...
namespace grid_sync { #define FIRST_UINT64_BIT ((uint64_t)1 << (sizeof(uint64_t) * 8 - 1)) template <typename T> __device__ T globalAsVolatile(volatile T& global_val) { return global_val; } template <bool X_BLOCK, bool Y_BLOCK, bool Z_BLOCK, bool PERSISTENT> __device__ void sync( int64_t& semaphore, cons...
namespace grid_sync { #define FIRST_UINT64_BIT ((uint64_t)1 << (sizeof(uint64_t) * 8 - 1)) template <typename T> __device__ T globalAsVolatile(volatile T& global_val) { return global_val; } template <bool X_BLOCK, bool Y_BLOCK, bool Z_BLOCK, bool PERSISTENT> __device__ void sync( int64_t& semaphore, const...
// !!! This is a file automatically generated by hipify!!! namespace index_utils { // Utility functions // Total size of provided dimension template <typename _dim3> __device__ __forceinline__ nvfuser_index_t size(const _dim3& d) { return (nvfuser_index_t)d.x * (nvfuser_index_t)d.y * (nvfuser_index_t)d.z; } // Li...
namespace index_utils { // Utility functions // Total size of provided dimension template <typename _dim3> __device__ __forceinline__ nvfuser_index_t size(const _dim3& d) { return (nvfuser_index_t)d.x * (nvfuser_index_t)d.y * (nvfuser_index_t)d.z; } // Linearized indexing of idx based on dim, if bool==false that ...
__device__ unsigned int mulhilo32( unsigned int a, unsigned int b, unsigned int* result_high) { *result_high = __umulhi(a, b); return a * b; } __device__ uint4 single_round(uint4 ctr, uint2 key) { constexpr unsigned long kPhiloxSA = 0xD2511F53; constexpr unsigned long kPhiloxSB = 0xCD9E8D57; unsigned int hi0; ...
__device__ unsigned int mulhilo32( unsigned int a, unsigned int b, unsigned int* result_high) { *result_high = __umulhi(a, b); return a * b; } __device__ uint4 single_round(uint4 ctr, uint2 key) { constexpr unsigned long kPhiloxSA = 0xD2511F53; constexpr unsigned long kPhiloxSB = 0xCD9E8D57; uns...
// !!! This is a file automatically generated by hipify!!! // Utility macro for this file #define DEVICE_INLINE __device__ inline // Utility class for 2D swizzle: template <typename index_t> struct IndexGeneric { const index_t x = 0, y = 0; DEVICE_INLINE IndexGeneric(index_t x_, index_t y_) : x(x_), y(y_) {} }; ...
// Utility macro for this file #define DEVICE_INLINE __device__ inline // Utility class for 2D swizzle: template <typename index_t> struct IndexGeneric { const index_t x = 0, y = 0; DEVICE_INLINE IndexGeneric(index_t x_, index_t y_) : x(x_), y(y_) {} }; // Default type for integration using Index2D = IndexGeneri...
// !!! This is a file automatically generated by hipify!!! template <typename T, int N> struct Tensor { __device__ T& operator[](nvfuser_index_t ind) { return data[ind]; }; T* data; nvfuser_index_t size[N]; nvfuser_index_t stride[N]; }; // Specialization for 0-dim case as it does not need size and stri...
template <typename T, int N> struct Tensor { __device__ T& operator[](nvfuser_index_t ind) { return data[ind]; }; T* data; nvfuser_index_t size[N]; nvfuser_index_t stride[N]; }; // Specialization for 0-dim case as it does not need size and stride arrays. // They will be an error as well since zero-leng...
// !!! This is a file automatically generated by hipify!!! // Type trait utils template <typename Type, bool is_volatile> struct MaybeVolatile; template <typename Type> struct MaybeVolatile<Type, true> { using type = volatile Type; }; template <typename Type> struct MaybeVolatile<Type, false> { using type = Type...
// Type trait utils template <typename Type, bool is_volatile> struct MaybeVolatile; template <typename Type> struct MaybeVolatile<Type, true> { using type = volatile Type; }; template <typename Type> struct MaybeVolatile<Type, false> { using type = Type; }; template <typename... Types> struct TypeList {}; tem...
// !!! This is a file automatically generated by hipify!!! namespace warp { template < bool SINGLE_WARP, typename T, typename Func, typename _dim3ti, typename _dim3bd> __device__ void warpReduceTIDX( T& out, const T& inp_val, Func reduction_op, const _dim3ti& thread_idx, const ...
namespace warp { template < bool SINGLE_WARP, typename T, typename Func, typename _dim3ti, typename _dim3bd> __device__ void warpReduceTIDX( T& out, const T& inp_val, Func reduction_op, const _dim3ti& thread_idx, const _dim3bd& block_dim, T* shared_mem, bool read_write_...
// !!! This is a file automatically generated by hipify!!! namespace warp { template < bool SINGLE_WARP, typename T, typename Func, typename _dim3ti, typename _dim3bd> __device__ void warpReduceTIDX( T& out, const T& inp_val, Func reduction_op, const _dim3ti& thread_idx, const ...
namespace warp { template < bool SINGLE_WARP, typename T, typename Func, typename _dim3ti, typename _dim3bd> __device__ void warpReduceTIDX( T& out, const T& inp_val, Func reduction_op, const _dim3ti& thread_idx, const _dim3bd& block_dim, T* shared_mem, bool read_write_...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: HIP on Windows requires that the enclos...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: CUDA on Windows requires that the enclosing function // of a __device__ lambda not have internal ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/hip\BinaryInternal.h> #include <c10/hip/HIPGuard.h> #include <c10/hip/HIPMathCompat.h> #i...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/BinaryInternal.h> #include <c10/cuda/CUDAGuard.h> #include <c10/cuda/CUDAMathCompat.h...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <c10/hip/HIPGuard.h> #include <c10/hip/HIPMa...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <c10/cuda/CUDAGuard.h> #include <c10/cuda/CUDAMathCompat.h> #include <c10/util/TypeSafeSignMath.h> #inc...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <c10/hip/HIPGuard.h> #include <c10/hip/HIPMa...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <c10/cuda/CUDAGuard.h> #include <c10/cuda/CUDAMathCompat.h> #include <c10/util/TypeSafeSignMath.h> #inc...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: HIP on Windows requires that the enclos...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: CUDA on Windows requires that the enclosing function // of a __device__ lambda not have internal ...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/hip\BinaryInternal.h> #include ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/BinaryInternal.h> #include <c10/cuda/CUDAGuard.h> #include <c10/cuda/CUDAMathCompat.h...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\cub.cuh> #include <ATen/hip\HIPConfig.h> namespace at { namespace hip { namespace cub { namespace { template <typename scalar_t> struct SumOp { __device__ scalar_t operator () (scalar_t a, scalar_t b) co...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/cub.cuh> #include <ATen/cuda/CUDAConfig.h> namespace at { namespace cuda { namespace cub { namespace { template <typename scalar_t> struct SumOp { __device__ scalar_t operator () (scalar_t a, scalar_t b) const { return a + b; } }; } template <typename in...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/TensorIterator.h> #include <c10/util/TypeSafeSignMath.h> #include ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/TensorIterator.h> #include <c10/util/TypeSafeSignMath.h> #include <type_traits> // NOTE: CUDA on Windows requires that the ...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: HIP on Windows requires that the enclos...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> // NOTE: CUDA on Windows requires that the enclosing function // of a __device__ lambda not have internal ...
#include "hip/hip_runtime.h" #pragma once #include <thrust/tuple.h> #include <ATen/native/SharedReduceOps.h> #include <ATen/hip\DeviceUtils.cuh> namespace at { namespace native { namespace hip_utils { constexpr int kHIPBlockReduceNumThreads = 512; constexpr int kHIPBlockReduceMaxThreads = C10_WARP_SIZE * C10_WARP...
#pragma once #include <thrust/tuple.h> #include <ATen/native/SharedReduceOps.h> #include <ATen/cuda/DeviceUtils.cuh> namespace at { namespace native { namespace cuda_utils { constexpr int kCUDABlockReduceNumThreads = 512; constexpr int kCUDABlockReduceMaxThreads = C10_WARP_SIZE * C10_WARP_SIZE; template <typen...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/hip\Loops.cuh> // NOTE: HIP on Windows requires that the enclo...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/Loops.cuh> // NOTE: CUDA on Windows requires that the enclosing function // of a __device__ lambda not have internal...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/TensorFactories.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/hip\Loops.cuh> // NOTE: HIP on Windows requires that the enclosing function // of a __...
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/TensorFactories.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/Loops.cuh> // NOTE: CUDA on Windows requires that the enclosing function // of a __device__ lambda not have internal linkage. namespace at:...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> #if defined(__HIPCC__) #include <hip/hip_runtime...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> #if defined(__CUDACC__) #include <cuda.h> #include <cuda_fp16.h> #include <c10/cuda/CUDAMathCompat.h> #eli...
// !!! This is a file automatically generated by hipify!!! #pragma once #if !defined(USE_ROCM) #include <hip/hip_runtime.h> // for TORCH_HIP_VERSION #endif #if !defined(USE_ROCM) #include <cub/version.cuh> #else #define CUB_VERSION 0 #endif // cub sort support for __nv_bfloat16 is added to cub 1.13 in: // https://...
#pragma once #if !defined(USE_ROCM) #include <cuda.h> // for CUDA_VERSION #endif #if !defined(USE_ROCM) #include <cub/version.cuh> #else #define CUB_VERSION 0 #endif // cub sort support for __nv_bfloat16 is added to cub 1.13 in: // https://github.com/NVIDIA/cub/pull/306 #if CUB_VERSION >= 101300 #define CUB_SUPPOR...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/core/Tensor.h> #include <ATen/Dispatch.h> #ifndef AT_PER_OPERATOR_HEADERS #include <ATen/NativeFunctions.h> #else #include <ATen/ops/_local_scalar_dense_native.h> #endif #include <ATen/hip\HIPContext....
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/core/Tensor.h> #include <ATen/Dispatch.h> #ifndef AT_PER_OPERATOR_HEADERS #include <ATen/NativeFunctions.h> #else #include <ATen/ops/_local_scalar_dense_native.h> #endif #include <ATen/cuda/CUDAContext.h> namespace at::native { Scalar _local_scalar_dense_cu...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/hip\ScanKernels.h> #include <ATen/native/hip\ScanUtils.cuh> #include <limits> #include <functional> namespace at::native { void launch_cumm...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/cuda/ScanKernels.h> #include <ATen/native/cuda/ScanUtils.cuh> #include <limits> #include <functional> namespace at::native { void launch_cummax_cuda_kernel(const TensorBase& self, const TensorBase& ...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/hip\ScanKernels.h> #include <ATen/native/hip\ScanUtils.cuh> namespace at::native { void launch_cumprod_hip_kernel(const TensorBase& result, ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/cuda/ScanKernels.h> #include <ATen/native/cuda/ScanUtils.cuh> namespace at::native { void launch_cumprod_cuda_kernel(const TensorBase& result, const TensorBase& self, int64_t dim) { AT_DISPATCH_ALL...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/hip\ScanKernels.h> #include <ATen/native/hip\ScanUtils.cuh> namespace at::native { void launch_cumsum_hip_kernel(const TensorBase& result, c...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/core/TensorBase.h> #include <ATen/Dispatch.h> #include <ATen/native/cuda/ScanKernels.h> #include <ATen/native/cuda/ScanUtils.cuh> namespace at::native { void launch_cumsum_cuda_kernel(const TensorBase& result, const TensorBase& self, int64_t dim) { AT_DISPATCH_ALL_...
// !!! This is a file automatically generated by hipify!!! #pragma once namespace at { namespace native { #if defined(USE_ROCM) // take these out when ROCm implements std:: math functions #include <math.h> template <typename scalar_t> static __forceinline__ __device__ scalar_t device_sqrt(scalar_t val); template <> ...
#pragma once namespace at { namespace native { #if defined(USE_ROCM) // take these out when ROCm implements std:: math functions #include <math.h> template <typename scalar_t> static __forceinline__ __device__ scalar_t device_sqrt(scalar_t val); template <> __forceinline__ __device__ float device_sqrt(float val) { ...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/hip\HIPApplyUtils.cuh> #include <ATen/AccumulateType.h> #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/cuda/CUDAApplyUtils.cuh> #include <ATen/AccumulateType.h> #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> #include <curand.h> #include <curand_kernel.h> #include <...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void cauchy_kernel(TensorIteratorBase& iter, double median, double sigma, c10...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void cauchy_kernel(TensorIteratorBase& iter, double median, double sigma, c10::optional<Generator> gen) { auto generator = get_gene...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void exponential_kernel(TensorIteratorBase& iter, double lambda, c10::optiona...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void exponential_kernel(TensorIteratorBase& iter, double lambda, c10::optional<Generator> gen) { auto generator = get_generator_or_...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void geometric_kernel(TensorIteratorBase& iter, double p_, c10::optional<Gene...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void geometric_kernel(TensorIteratorBase& iter, double p_, c10::optional<Generator> gen) { auto generator = get_generator_or_defaul...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void log_normal_kernel(TensorIteratorBase& iter, double mean, double std, c10...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void log_normal_kernel(TensorIteratorBase& iter, double mean, double std, c10::optional<Generator> gen) { auto generator = get_gene...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/hip\HIPEvent.h> #include <ATen/hip/detail\UnpackRaw.cuh> #include <ATen/hip/detail\HIPHooks.h> #include <ATen/detail/HIPHooksInterface.h> #include <c10/core/StreamGuard.h> #include <c10/hip/HI...
#pragma once #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/cuda/CUDAEvent.h> #include <ATen/cuda/detail/UnpackRaw.cuh> #include <ATen/cuda/detail/CUDAHooks.h> #include <ATen/detail/CUDAHooksInterface.h> #include <c10/core/StreamGuard.h> #include <c10/cuda/CUDAGraphsC10Utils.h> #include <c10/cuda/CUDAGuard.h...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void normal_kernel(const TensorBase &self, double mean, double std, c10::opti...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void normal_kernel(const TensorBase &self, double mean, double std, c10::optional<Generator> gen) { auto generator = get_generator_...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void random_from_to_kernel(TensorIteratorBase& iter, uint64_t range, int64_t ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void random_from_to_kernel(TensorIteratorBase& iter, uint64_t range, int64_t base, c10::optional<Generator> gen_) { auto gen = get_...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/hip\HIPGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/hip\DistributionTemplates.h> namespace at::native { void uniform_kernel(TensorIteratorBase& iter, double from, double to, c10::op...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/cuda/CUDAGeneratorImpl.h> #include <ATen/native/UnaryOps.h> #include <ATen/native/cuda/DistributionTemplates.h> namespace at::native { void uniform_kernel(TensorIteratorBase& iter, double from, double to, c10::optional<Generator> gen) { auto generator = get_generato...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/core/Tensor.h> #include <ATen/hip\Atomic.cuh> #include <ATen/hip\HIPContext.h> #include <ATen/TensorUtils.h> namespace at { namespace native { Tensor embedding_backward_hip_kernel( const Tensor &grad, const Tensor &orig_in...
#pragma once #include <ATen/core/Tensor.h> #include <ATen/cuda/Atomic.cuh> #include <ATen/cuda/CUDAContext.h> #include <ATen/TensorUtils.h> namespace at { namespace native { Tensor embedding_backward_cuda_kernel( const Tensor &grad, const Tensor &orig_indices, const Tensor &sorted_indices, const Tens...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/Fill.h> #include <c10/core/Scalar.h> namespace at::native { ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/DispatchStub.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/Fill.h> #include <c10/core/Scalar.h> namespace at::native { template<typename scalar_t> struct FillFunctor { FillFun...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/native/sparse/SparseStubs.h> #include <ATen/native/sparse/FlattenIndicesCommon.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/hip\KernelUtils.cuh> #include <ATen/hip/detail\OffsetCalculat...
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/native/sparse/SparseStubs.h> #include <ATen/native/sparse/FlattenIndicesCommon.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/cuda/KernelUtils.cuh> #include <ATen/cuda/detail/OffsetCalculator.cuh> #include <ATen/AccumulateType.h> namespace at::...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/NumericUtils.h> namespace at::native { // std:: does not have clamp functors template <typename T> struct minimum { __device__ T operator()(const T& a, const T& b) const { return (_isnan(a) || a < b) ? a : b; } }; templa...
#pragma once #include <ATen/NumericUtils.h> namespace at::native { // std:: does not have clamp functors template <typename T> struct minimum { __device__ T operator()(const T& a, const T& b) const { return (_isnan(a) || a < b) ? a : b; } }; template <typename T> struct maximum { __device__ T operator()(...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/TypeDefault.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adam_amsgrad_impl.cuh> #include <ATen/native/hip\fused_adam_impl.cuh> #include <c10/util/Exception.h> namespace at:...
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/TypeDefault.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adam_amsgrad_impl.cuh> #include <ATen/native/cuda/fused_adam_impl.cuh> #include <c10/util/Exception.h> namespace at::native { // note(crcrpar): To observe the CI rules, i.e...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/TypeDefault.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adamw_amsgrad_impl.cuh> #include <ATen/native/hip\fused_adamw_impl.cuh> #include <c10/util/Exception.h> namespace a...
#define TORCH_ASSERT_ONLY_METHOD_OPERATORS #include <ATen/TypeDefault.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adamw_amsgrad_impl.cuh> #include <ATen/native/cuda/fused_adamw_impl.cuh> #include <c10/util/Exception.h> namespace at { namespace native { // note(crcrpar): To observe the ...
// !!! This is a file automatically generated by hipify!!! #include <ATen/native/hip\fused_adamw_amsgrad_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adam_utils.cuh> #include <ATen/native/hip\MultiTensorApply.cuh> #include <vector> namespace at { namespac...
#include <ATen/native/cuda/fused_adamw_amsgrad_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adam_utils.cuh> #include <ATen/native/cuda/MultiTensorApply.cuh> #include <vector> namespace at { namespace native { void _fused_adamw_amsgrad_cuda_impl_( at...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/Tensor.h> #include <c10/util/Half.h> #include <hip/hip_runtime.h> #include <hip/hip_runtime.h> #include <hip/hip_fp16.h> namespace at { template <> inline __half* Tensor::data() const { return reinterpret_cast<__half*>(data<Hal...
#pragma once #include <ATen/Tensor.h> #include <c10/util/Half.h> #include <cuda.h> #include <cuda_runtime.h> #include <cuda_fp16.h> namespace at { template <> inline __half* Tensor::data() const { return reinterpret_cast<__half*>(data<Half>()); } } // namespace at ###
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adamw_amsgrad_hip_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg...
#pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adamw_amsgrad_cuda_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, const double lr,...
// !!! This is a file automatically generated by hipify!!! #include <ATen/native/hip\fused_adamw_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adam_utils.cuh> #include <ATen/native/hip\MultiTensorApply.cuh> #include <vector> namespace at { namespace native...
#include <ATen/native/cuda/fused_adamw_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adam_utils.cuh> #include <ATen/native/cuda/MultiTensorApply.cuh> #include <vector> namespace at { namespace native { void _fused_adamw_cuda_impl_( at::TensorList par...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adamw_hip_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList state_steps, co...
#pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adamw_cuda_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList state_steps, const double lr, const double beta1, const double be...
// !!! This is a file automatically generated by hipify!!! #include <ATen/native/hip\fused_adam_amsgrad_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adam_utils.cuh> #include <ATen/native/hip\MultiTensorApply.cuh> #include <vector> namespace at::native { ...
#include <ATen/native/cuda/fused_adam_amsgrad_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adam_utils.cuh> #include <ATen/native/cuda/MultiTensorApply.cuh> #include <vector> namespace at::native { void _fused_adam_amsgrad_cuda_impl_( at::TensorList ...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adam_amsgrad_hip_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_...
#pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adam_amsgrad_cuda_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList max_exp_avg_sqs, at::TensorList state_steps, const double lr, ...
// !!! This is a file automatically generated by hipify!!! #include <ATen/native/hip\fused_adam_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/hip\fused_adam_utils.cuh> #include <ATen/native/hip\MultiTensorApply.cuh> #include <vector> namespace at::native { void _fu...
#include <ATen/native/cuda/fused_adam_impl.cuh> #include <ATen/Dispatch.h> #include <ATen/native/ForeachUtils.h> #include <ATen/native/cuda/fused_adam_utils.cuh> #include <ATen/native/cuda/MultiTensorApply.cuh> #include <vector> namespace at::native { void _fused_adam_cuda_impl_( at::TensorList params, at::...
// !!! This is a file automatically generated by hipify!!! #pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adam_hip_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList state_steps, con...
#pragma once #include <ATen/core/Tensor.h> namespace at { namespace native { void _fused_adam_cuda_impl_( at::TensorList params, at::TensorList grads, at::TensorList exp_avgs, at::TensorList exp_avg_sqs, at::TensorList state_steps, const double lr, const double beta1, const double bet...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/hip\Math.cuh> #include <ATen/native/TensorIterator.h> #include...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> #include <ATen/native/cuda/ji...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #include "hip/hip_runtime.h" #pragma once #include <ATen/ceil_div.h> #include <ATen/hip\DeviceUtils.cuh> #include <ATen/hip\AsmUtils.cuh> #include <c10/macros/Macros.h> // Collection of in-kernel scan / prefix sum utilities namespace at { namespace hip { /...
#pragma once #include <ATen/ceil_div.h> #include <ATen/cuda/DeviceUtils.cuh> #include <ATen/cuda/AsmUtils.cuh> #include <c10/macros/Macros.h> // Collection of in-kernel scan / prefix sum utilities namespace at { namespace cuda { // Inclusive prefix sum for binary vars using intra-warp voting + // shared memory tem...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/hip\JitLoops.cuh> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/hip\Math.cuh> #include <ATen/nativ...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/BinaryOps.h> #include <ATen/native/Math.h> #include <ATen/native/cuda/Math.cuh> #include <ATen/native/cuda/jit_utils.h> namespace at::native { name...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/hip\Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> #include <ATen/OpMathType.h> #include <c10/util/M...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/cuda/Loops.cuh> #include <ATen/native/TensorIterator.h> #include <ATen/native/BinaryOps.h> #include <ATen/OpMathType.h> #include <c10/util/MathConstants.h> // NOTE: CUDA on Windows requires that th...
// !!! This is a file automatically generated by hipify!!! #define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #in...
#define TORCH_ASSERT_NO_OPERATORS #include <ATen/native/UnaryOps.h> #include <limits> #include <ATen/AccumulateType.h> #include <ATen/Dispatch.h> #include <ATen/native/DispatchStub.h> #include <ATen/native/Math.h> #include <ATen/native/TensorIterator.h> #include <ATen/native/cuda/JitLoops.cuh> #include <ATen/native...