system stringclasses 7
values | input stringlengths 10 94.5k | output stringlengths 14 55.5k |
|---|---|---|
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0004e1ff_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "kernel.hip"
.globl _Z26__device_stub__naiveKernelPfS_i # -- Begin function _Z26__device_stub__naiveKernelPfS_i
.p2align 4, 0x90
.type _Z26__device_stub__naiveKernelPfS_i,@function
_Z26__device_stub__naiveKernelPfS_i: # @_Z26__device_stub__naiveKernelPfS_i
.cfi_startproc
# %bb.0:
subq $104, %rsp
.cfi_def... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
#include <cuda.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
cudaDeviceSynchronize();
}
return 0;
} | code for sm_80
Function : _Z1Ki
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R2, SR_LANEID ; /* 0x0000000000027919 */
/* 0x000e220000000000 */
/*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include <cuda.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
cudaDeviceSynchronize();
}
return 0;
} | .file "tmpxft_00133e93_00000000-6_nvprof3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#include <cuda.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
cudaDeviceSynchronize();
}
return 0;
} | #include <stdio.h>
#include <hip/hip_runtime.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
hipDeviceSynchronize();
}
return 0;
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <stdio.h>
#include <hip/hip_runtime.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
hipDeviceSynchronize();
}
return 0;
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z1Ki
.globl _Z1Ki
.p2align 8
.type _Z1Ki,@function
_Z1Ki:
s_load_b32 s0, s[0:1], 0x0
s_mov_b32 s2, exec_lo
s_mov_b32 s1, exec_lo
v_mbcnt_lo_u32_b32 v0, s2, 0
s_waitcnt lgkmcnt(0)
s_lshl_b32 s0, s0, 1
s_del... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <stdio.h>
#include <hip/hip_runtime.h>
__device__ int sumg = 0;
__global__ void K(int num) {
num += num;
++num;
atomicAdd(&sumg, num);
__shared__ int sum;
sum = 0;
__syncthreads();
sum += num;
}
int main() {
for (unsigned ii = 0; ii < 100; ++ii) {
K<<<5, 32>>>(ii);
hipDeviceSynchronize();
}
return 0;
} | .text
.file "nvprof3.hip"
.globl _Z16__device_stub__Ki # -- Begin function _Z16__device_stub__Ki
.p2align 4, 0x90
.type _Z16__device_stub__Ki,@function
_Z16__device_stub__Ki: # @_Z16__device_stub__Ki
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_cfa_offset 80
movl %edi, 12(%rsp)
leaq 12(%rs... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z1Ki
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R2, SR_LANEID ; /* 0x0000000000027919 */
/* 0x000e220000000000 */
/*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z1Ki
.globl _Z1Ki
.p2align 8
.type _Z1Ki,@function
_Z1Ki:
s_load_b32 s0, s[0:1], 0x0
s_mov_b32 s2, exec_lo
s_mov_b32 s1, exec_lo
v_mbcnt_lo_u32_b32 v0, s2, 0
s_waitcnt lgkmcnt(0)
s_lshl_b32 s0, s0, 1
s_del... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00133e93_00000000-6_nvprof3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... | .text
.file "nvprof3.hip"
.globl _Z16__device_stub__Ki # -- Begin function _Z16__device_stub__Ki
.p2align 4, 0x90
.type _Z16__device_stub__Ki,@function
_Z16__device_stub__Ki: # @_Z16__device_stub__Ki
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_cfa_offset 80
movl %edi, 12(%rsp)
leaq 12(%rs... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /*
* Copyright (c) 2019, NVIDIA CORPORATION.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to ... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
* Copyright (c) 2019, NVIDIA CORPORATION.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to ... | /*
* Copyright (c) 2019, NVIDIA CORPORATION.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /*
* Copyright (c) 2019, NVIDIA CORPORATION.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to ... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | extern "C"
__global__ void initDbIndexKernel(int totalVars, int totalPreds, int *d_varDomainSizes,
int *d_predBaseIdx, int *d_predVarMat, int *d_dbIndex,
long totalGroundings)
{
long idx = blockIdx.x * blockDim.x + threadIdx.x;
if(idx < totalGroundings)
{
long baseDbIndex = idx * totalPreds;
for(int i = 0; i < totalPre... | .file "tmpxft_000829c5_00000000-6_mlnCudaKernels.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | extern "C"
__global__ void initDbIndexKernel(int totalVars, int totalPreds, int *d_varDomainSizes,
int *d_predBaseIdx, int *d_predVarMat, int *d_dbIndex,
long totalGroundings)
{
long idx = blockIdx.x * blockDim.x + threadIdx.x;
if(idx < totalGroundings)
{
long baseDbIndex = idx * totalPreds;
for(int i = 0; i < totalPre... | #include <hip/hip_runtime.h>
extern "C"
__global__ void initDbIndexKernel(int totalVars, int totalPreds, int *d_varDomainSizes,
int *d_predBaseIdx, int *d_predVarMat, int *d_dbIndex,
long totalGroundings)
{
long idx = blockIdx.x * blockDim.x + threadIdx.x;
if(idx < totalGroundings)
{
long baseDbIndex = idx * totalPreds... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
extern "C"
__global__ void initDbIndexKernel(int totalVars, int totalPreds, int *d_varDomainSizes,
int *d_predBaseIdx, int *d_predVarMat, int *d_dbIndex,
long totalGroundings)
{
long idx = blockIdx.x * blockDim.x + threadIdx.x;
if(idx < totalGroundings)
{
long baseDbIndex = idx * totalPreds... | .text
.file "mlnCudaKernels.hip"
.globl __device_stub__initDbIndexKernel # -- Begin function __device_stub__initDbIndexKernel
.p2align 4, 0x90
.type __device_stub__initDbIndexKernel,@function
__device_stub__initDbIndexKernel: # @__device_stub__initDbIndexKernel
.cfi_startproc
# %bb.0:
subq $152, %rsp
.cfi_def_cfa... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000829c5_00000000-6_mlnCudaKernels.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... | .text
.file "mlnCudaKernels.hip"
.globl __device_stub__initDbIndexKernel # -- Begin function __device_stub__initDbIndexKernel
.p2align 4, 0x90
.type __device_stub__initDbIndexKernel,@function
__device_stub__initDbIndexKernel: # @__device_stub__initDbIndexKernel
.cfi_startproc
# %bb.0:
subq $152, %rsp
.cfi_def_cfa... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <cuda.h>
#include <cuda_runtime.h>
///all parallel implementations of this algorithim will require two functions or else delay a function significantly
__global__
void sundPartOnePerRow(int bound, bool * findArray)
{
int idx = blockDim.x * blockIdx.x + threadIdx.x;
if(idx < 1)
{
return;
}
if (idx > bound)
re... | .file "tmpxft_00186a25_00000000-6_parallelSieve.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <cuda.h>
#include <cuda_runtime.h>
///all parallel implementations of this algorithim will require two functions or else delay a function significantly
__global__
void sundPartOnePerRow(int bound, bool * findArray)
{
int idx = blockDim.x * blockIdx.x + threadIdx.x;
if(idx < 1)
{
return;
}
if (idx > bound)
re... | #include <hip/hip_runtime.h>
///all parallel implementations of this algorithim will require two functions or else delay a function significantly
__global__
void sundPartOnePerRow(int bound, bool * findArray)
{
int idx = blockDim.x * blockIdx.x + threadIdx.x;
if(idx < 1)
{
return;
}
if (idx > bound)
return;
int deno... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
///all parallel implementations of this algorithim will require two functions or else delay a function significantly
__global__
void sundPartOnePerRow(int bound, bool * findArray)
{
int idx = blockDim.x * blockIdx.x + threadIdx.x;
if(idx < 1)
{
return;
}
if (idx > bound)
return;
int deno... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z17sundPartOnePerRowiPb
.globl _Z17sundPartOnePerRowiPb
.p2align 8
.type _Z17sundPartOnePerRowiPb,@function
_Z17sundPartOnePerRowiPb:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x1c
s_load_b32 s3, s[0:1], 0x0
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
///all parallel implementations of this algorithim will require two functions or else delay a function significantly
__global__
void sundPartOnePerRow(int bound, bool * findArray)
{
int idx = blockDim.x * blockIdx.x + threadIdx.x;
if(idx < 1)
{
return;
}
if (idx > bound)
return;
int deno... | .text
.file "parallelSieve.hip"
.globl _Z32__device_stub__sundPartOnePerRowiPb # -- Begin function _Z32__device_stub__sundPartOnePerRowiPb
.p2align 4, 0x90
.type _Z32__device_stub__sundPartOnePerRowiPb,@function
_Z32__device_stub__sundPartOnePerRowiPb: # @_Z32__device_stub__sundPartOnePerRowiPb
.cfi_startproc
# %bb.0:
... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00186a25_00000000-6_parallelSieve.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... | .text
.file "parallelSieve.hip"
.globl _Z32__device_stub__sundPartOnePerRowiPb # -- Begin function _Z32__device_stub__sundPartOnePerRowiPb
.p2align 4, 0x90
.type _Z32__device_stub__sundPartOnePerRowiPb,@function
_Z32__device_stub__sundPartOnePerRowiPb: # @_Z32__device_stub__sundPartOnePerRowiPb
.cfi_startproc
# %bb.0:
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + threadIdx.x;
sdata[tid] = g_ida... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + threadIdx.x;
sdata[tid] = g_ida... | .file "tmpxft_0006e232_00000000-6_CU_tool.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2039:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + threadIdx.x;
sdata[tid] = g_ida... | #include <hip/hip_runtime.h>
#ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + th... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + th... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#ifndef BLOCK_SIZE
#define BLOCK_SIZE 512
#endif
template <class DataType>
__global__ void reduce(DataType *g_idata, DataType *g_odata,int N) {
//extern __shared__ int sdata[];
__shared__ DataType sdata[BLOCK_SIZE];
unsigned int tid = threadIdx.x;
unsigned int i = blockIdx.x*blockDim.x + th... | .text
.file "CU_tool.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project ... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0006e232_00000000-6_CU_tool.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2039:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... | .text
.file "CU_tool.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include <cuda_runtime.h>
//#include <cutil.h>
#define TILE_WIDTH 64
#define WIDTH_PER_THREAD 4
#define SW TILE_WIDTH/WIDTH_PER_THREAD
#define N 2048
void err_handling(cudaError_t *err, const char *str)
{
if (*err != cudaSuccess) {
printf("%s\n", str);
exit(EXIT_FAILURE);
}
}
__global__ void matMul(c... | .file "tmpxft_000d37cd_00000000-6_tiled_matmul-3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#include <cuda_runtime.h>
//#include <cutil.h>
#define TILE_WIDTH 64
#define WIDTH_PER_THREAD 4
#define SW TILE_WIDTH/WIDTH_PER_THREAD
#define N 2048
void err_handling(cudaError_t *err, const char *str)
{
if (*err != cudaSuccess) {
printf("%s\n", str);
exit(EXIT_FAILURE);
}
}
__global__ void matMul(c... | #include <stdio.h>
#include <hip/hip_runtime.h>
//#include <cutil.h>
#define TILE_WIDTH 64
#define WIDTH_PER_THREAD 4
#define SW TILE_WIDTH/WIDTH_PER_THREAD
#define N 2048
void err_handling(hipError_t *err, const char *str)
{
if (*err != hipSuccess) {
printf("%s\n", str);
exit(EXIT_FAILURE);
}
}
__global__ void matMul(... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <stdio.h>
#include <hip/hip_runtime.h>
//#include <cutil.h>
#define TILE_WIDTH 64
#define WIDTH_PER_THREAD 4
#define SW TILE_WIDTH/WIDTH_PER_THREAD
#define N 2048
void err_handling(hipError_t *err, const char *str)
{
if (*err != hipSuccess) {
printf("%s\n", str);
exit(EXIT_FAILURE);
}
}
__global__ void matMul(... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6matMulPKfS0_Pfiii
.globl _Z6matMulPKfS0_Pfiii
.p2align 8
.type _Z6matMulPKfS0_Pfiii,@function
_Z6matMulPKfS0_Pfiii:
s_load_b32 s2, s[0:1], 0x1c
v_bfe_u32 v22, v0, 10, 10
v_and_b32_e32 v23, 0x3ff, v0
s_delay_alu instid0(VALU_DEP_2) | ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <stdio.h>
#include <hip/hip_runtime.h>
//#include <cutil.h>
#define TILE_WIDTH 64
#define WIDTH_PER_THREAD 4
#define SW TILE_WIDTH/WIDTH_PER_THREAD
#define N 2048
void err_handling(hipError_t *err, const char *str)
{
if (*err != hipSuccess) {
printf("%s\n", str);
exit(EXIT_FAILURE);
}
}
__global__ void matMul(... | .text
.file "tiled_matmul-3.hip"
.globl _Z12err_handlingP10hipError_tPKc # -- Begin function _Z12err_handlingP10hipError_tPKc
.p2align 4, 0x90
.type _Z12err_handlingP10hipError_tPKc,@function
_Z12err_handlingP10hipError_tPKc: # @_Z12err_handlingP10hipError_tPKc
.cfi_startproc
# %bb.0:
cmpl $0, (%rdi)
jne .LBB0_2
... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000d37cd_00000000-6_tiled_matmul-3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... | .text
.file "tiled_matmul-3.hip"
.globl _Z12err_handlingP10hipError_tPKc # -- Begin function _Z12err_handlingP10hipError_tPKc
.p2align 4, 0x90
.type _Z12err_handlingP10hipError_tPKc,@function
_Z12err_handlingP10hipError_tPKc: # @_Z12err_handlingP10hipError_tPKc
.cfi_startproc
# %bb.0:
cmpl $0, (%rdi)
jne .LBB0_2
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} | code for sm_80
Function : _Z3addPfS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R3, SR_CTAID.Y ; /* 0x0000000000037919 */
/* 0x000e280000002600 */
/*0020*/ S2... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} | .file "tmpxft_000babb0_00000000-6_add.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPfS_ii
.globl _Z3addPfS_ii
.p2align 8
.type _Z3addPfS_ii,@function
_Z3addPfS_ii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x10
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 v3, v0, 10, 10
s_waitc... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void add(float *a, float *b, int nx, int ny)
{
int x = blockIdx.x * blockDim.x + threadIdx.x;
int y = blockIdx.y * blockDim.y + threadIdx.y;
int idx = x + y*nx;
if (x<nx && y<ny) b[idx] += a[idx] * .125;
} | .text
.file "add.hip"
.globl _Z18__device_stub__addPfS_ii # -- Begin function _Z18__device_stub__addPfS_ii
.p2align 4, 0x90
.type _Z18__device_stub__addPfS_ii,@function
_Z18__device_stub__addPfS_ii: # @_Z18__device_stub__addPfS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset 128
movq %rdi, 7... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z3addPfS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R3, SR_CTAID.Y ; /* 0x0000000000037919 */
/* 0x000e280000002600 */
/*0020*/ S2... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPfS_ii
.globl _Z3addPfS_ii
.p2align 8
.type _Z3addPfS_ii,@function
_Z3addPfS_ii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x10
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 v3, v0, 10, 10
s_waitc... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000babb0_00000000-6_add.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... | .text
.file "add.hip"
.globl _Z18__device_stub__addPfS_ii # -- Begin function _Z18__device_stub__addPfS_ii
.p2align 4, 0x90
.type _Z18__device_stub__addPfS_ii,@function
_Z18__device_stub__addPfS_ii: # @_Z18__device_stub__addPfS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset 128
movq %rdi, 7... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | //Based on the work of Andrew Krepps
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(int * a, int * b, int * c)
{
... | code for sm_80
Function : _Z3modPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2200000025... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | //Based on the work of Andrew Krepps
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(int * a, int * b, int * c)
{
... | .file "tmpxft_000f8245_00000000-6_assignment.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB4757:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cud... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | //Based on the work of Andrew Krepps
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(int * a, int * b, int * c)
{
... | //Based on the work of Andrew Krepps
#include <hip/hip_runtime.h>
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | //Based on the work of Andrew Krepps
#include <hip/hip_runtime.h>
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPiS_S_
.globl _Z3addPiS_S_
.p2align 8
.type _Z3addPiS_S_,@function
_Z3addPiS_S_:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b128 s[4:7], s[0:1], 0x0
s_load_b64 s[0:1], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | //Based on the work of Andrew Krepps
#include <hip/hip_runtime.h>
#include <iostream>
#include <random>
#include <stdio.h>
__global__ void add(int * a, int * b, int * c)
{
const unsigned int thread_idx = (blockIdx.x * blockDim.x) + threadIdx.x;
c[thread_idx] = a[thread_idx] + b[thread_idx];
}
__global__ void subtract(... | .text
.file "assignment.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z18__device_stub__addPiS_S_ # -- Begin function _Z18__device_stub__addPiS_S_
.p2align 4, 0x9... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z3modPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2200000025... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPiS_S_
.globl _Z3addPiS_S_
.p2align 8
.type _Z3addPiS_S_,@function
_Z3addPiS_S_:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b128 s[4:7], s[0:1], 0x0
s_load_b64 s[0:1], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... | code for sm_80
Function : _Z10matrix_mulPiS_S_iii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e280000002500 */
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... | .file "tmpxft_000a1782_00000000-6_cuda_matrix_global.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10matrix_mulPiS_S_iii
.globl _Z10matrix_mulPiS_S_iii
.p2align 8
.type _Z10matrix_mulPiS_S_iii,@function
_Z10matrix_mulPiS_S_iii:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x34
s_load_b64 s[2:3], s[0:1], 0x1c
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /*
* Alexandre Maros - 2016
*
* Cuda Matrix Multiplication with Global Memory.
*
* nvcc cuda_matrix_global.cu -o cg.o
*
* Implemented by Alexandre Maros for learning purposes.
* A version of this code using Shared Memory is in here:
* https://github.com/alepmaros/cuda_matrix_multiplication
*
* Distributed under the MIT... | .text
.file "cuda_matrix_global.hip"
.globl _Z25__device_stub__matrix_mulPiS_S_iii # -- Begin function _Z25__device_stub__matrix_mulPiS_S_iii
.p2align 4, 0x90
.type _Z25__device_stub__matrix_mulPiS_S_iii,@function
_Z25__device_stub__matrix_mulPiS_S_iii: # @_Z25__device_stub__matrix_mulPiS_S_iii
.cfi_startproc
# %bb.0:
... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z10matrix_mulPiS_S_iii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e280000002500 */
... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10matrix_mulPiS_S_iii
.globl _Z10matrix_mulPiS_S_iii
.p2align 8
.type _Z10matrix_mulPiS_S_iii,@function
_Z10matrix_mulPiS_S_iii:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x34
s_load_b64 s[2:3], s[0:1], 0x1c
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 ... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000a1782_00000000-6_cuda_matrix_global.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... | .text
.file "cuda_matrix_global.hip"
.globl _Z25__device_stub__matrix_mulPiS_S_iii # -- Begin function _Z25__device_stub__matrix_mulPiS_S_iii
.p2align 4, 0x90
.type _Z25__device_stub__matrix_mulPiS_S_iii,@function
_Z25__device_stub__matrix_mulPiS_S_iii: # @_Z25__device_stub__matrix_mulPiS_S_iii
.cfi_startproc
# %bb.0:
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (int*)malloc(size);
hc = (int*)... | code for sm_80
Function : _Z13kernel_vecaddPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_TID.X ; /* 0x0000000000067919 */
/* 0x000e220000002100 */
/*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (int*)malloc(size);
hc = (int*)... | .file "tmpxft_000b408c_00000000-6_vecadd.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (int*)malloc(size);
hc = (int*)... | #include <hip/hip_runtime.h>
#include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (i... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (i... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z13kernel_vecaddPiS_S_
.globl _Z13kernel_vecaddPiS_S_
.p2align 8
.type _Z13kernel_vecaddPiS_S_,@function
_Z13kernel_vecaddPiS_S_:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b128 s[4:7], s[0:1], 0x0
s_load_b64 s[0:1], s[0:1], 0x10
s_wait... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#define ELEMS 20
__global__ void kernel_vecadd(int *a, int *b, int *c){
int index = threadIdx.x + blockIdx.x * blockDim.x;
c[index] = a[index] + b[index];
}
int main(){
int *ha,*hb,*hc;
int *da,*db,*dc;
int i;
int size = ELEMS*sizeof(int);
ha = (int*)malloc(size);
hb = (i... | .text
.file "vecadd.hip"
.globl _Z28__device_stub__kernel_vecaddPiS_S_ # -- Begin function _Z28__device_stub__kernel_vecaddPiS_S_
.p2align 4, 0x90
.type _Z28__device_stub__kernel_vecaddPiS_S_,@function
_Z28__device_stub__kernel_vecaddPiS_S_: # @_Z28__device_stub__kernel_vecaddPiS_S_
.cfi_startproc
# %bb.0:
subq $104, %... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z13kernel_vecaddPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_TID.X ; /* 0x0000000000067919 */
/* 0x000e220000002100 */
/*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z13kernel_vecaddPiS_S_
.globl _Z13kernel_vecaddPiS_S_
.p2align 8
.type _Z13kernel_vecaddPiS_S_,@function
_Z13kernel_vecaddPiS_S_:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b128 s[4:7], s[0:1], 0x0
s_load_b64 s[0:1], s[0:1], 0x10
s_wait... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000b408c_00000000-6_vecadd.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "vecadd.hip"
.globl _Z28__device_stub__kernel_vecaddPiS_S_ # -- Begin function _Z28__device_stub__kernel_vecaddPiS_S_
.p2align 4, 0x90
.type _Z28__device_stub__kernel_vecaddPiS_S_,@function
_Z28__device_stub__kernel_vecaddPiS_S_: # @_Z28__device_stub__kernel_vecaddPiS_S_
.cfi_startproc
# %bb.0:
subq $104, %... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | // numInterior: (NfIn)
// intrplWgts: (NfIn*maxK, 3)
// input: (NfIn*maxK, C)
// filter: (3, C, r)
// output: (NfIn, C*r)
__global__ void facet2facet_conv3d_forward(int NfIn, int C, int r, const int* numInterior,
const float* intrplWgts, const float* input,
const float* filter, float* output)
{
int idx = blockIdx.x*blo... | #include <hip/hip_runtime.h>
// numInterior: (NfIn)
// intrplWgts: (NfIn*maxK, 3)
// input: (NfIn*maxK, C)
// filter: (3, C, r)
// output: (NfIn, C*r)
__global__ void facet2facet_conv3d_forward(int NfIn, int C, int r, const int* numInterior,
const float* intrplWgts, const float* input,
const float* filter, float* outpu... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /**
* Copyright © 2018 - 2019 Sergei Iurevich Filippov, All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required... | .file "tmpxft_000c1027_00000000-6_kernels.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2036:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /**
* Copyright © 2018 - 2019 Sergei Iurevich Filippov, All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required... | #include <hip/hip_runtime.h>
/**
* Copyright © 2018 - 2019 Sergei Iurevich Filippov, All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LI... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
/**
* Copyright © 2018 - 2019 Sergei Iurevich Filippov, All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LI... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12kernel_scalePvffm
.globl _Z12kernel_scalePvffm
.p2align 8
.type _Z12kernel_scalePvffm,@function
_Z12kernel_scalePvffm:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x24
s_load_b64 s[2:3], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s4, s4, 0xf... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
/**
* Copyright © 2018 - 2019 Sergei Iurevich Filippov, All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LI... | .text
.file "kernels.hip"
.globl _Z27__device_stub__kernel_scalePvffm # -- Begin function _Z27__device_stub__kernel_scalePvffm
.p2align 4, 0x90
.type _Z27__device_stub__kernel_scalePvffm,@function
_Z27__device_stub__kernel_scalePvffm: # @_Z27__device_stub__kernel_scalePvffm
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cf... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000c1027_00000000-6_kernels.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2036:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... | .text
.file "kernels.hip"
.globl _Z27__device_stub__kernel_scalePvffm # -- Begin function _Z27__device_stub__kernel_scalePvffm
.p2align 4, 0x90
.type _Z27__device_stub__kernel_scalePvffm,@function
_Z27__device_stub__kernel_scalePvffm: # @_Z27__device_stub__kernel_scalePvffm
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cf... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0_out, float * * uz_3_0_out, f... | code for sm_80
Function : _Z10initializePfS_S_S_fffiii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ I2F.U32.RP R0, c[0x0][0x8] ; /* 0x0000020000007b06 */
/* 0x000e220... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0_out, float * * uz_3_0_out, f... | .file "tmpxft_000eda76_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0_out, float * * uz_3_0_out, f... | #include <hip/hip_runtime.h>
#define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z8gradientPPfS0_S0_S_S_S_S_fffiii
.globl _Z8gradientPPfS0_S0_S_S_S_S_fffiii
.p2align 8
.type _Z8gradientPPfS0_S0_S_S_S_S_fffiii,@function
_Z8gradientPPfS0_S0_S_S_S_S_fffiii:
s_clause 0x1
s_load_b64 s[2:3], s[0:1], 0x5c
s_load_b64 s[12:13], s[0:1], 0x48... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#define t_max 1
#define t 1
/*
(ux[0][0][0][0][1]=(alpha*(u[1][0][0][0][0]+u[-1][0][0][0][0])))
(uy[0][0][0][0][2]=(beta*(u[0][1][0][0][0]+u[0][-1][0][0][0])))
(uz[0][0][0][0][3]=(gamma*(u[0][0][1][0][0]+u[0][0][-1][0][0])))
*/
__global__ void gradient(float * * ux_1_0_out, float * * uy_2_0... | .text
.file "kernel.hip"
.globl _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii # -- Begin function _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii
.p2align 4, 0x90
.type _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii,@function
_Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii: # @_Z23__device_stub__gradient... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000eda76_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "kernel.hip"
.globl _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii # -- Begin function _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii
.p2align 4, 0x90
.type _Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii,@function
_Z23__device_stub__gradientPPfS0_S0_S_S_S_S_fffiii: # @_Z23__device_stub__gradient... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
* =====================================================================================
*
* Filename: lud.cu
*
* Description: The main wrapper for the suite
*
* Version: 1.0
* Created: 10/22/2009 08:40:34 PM
* Revision: none
* Compiler: gcc
*
* Author: Liang Wang (lw2aw), lw2aw@virginia.edu
* Company: CS@UVa
*
* ===... | .file "tmpxft_00099cbc_00000000-6_lud.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2085:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
* =====================================================================================
*
* Filename: lud.cu
*
* Description: The main wrapper for the suite
*
* Version: 1.0
* Created: 10/22/2009 08:40:34 PM
* Revision: none
* Compiler: gcc
*
* Author: Liang Wang (lw2aw), lw2aw@virginia.edu
* Company: CS@UVa
*
* ===... | /*
* =====================================================================================
*
* Filename: lud.cu
*
* Description: The main wrapper for the suite
*
* Version: 1.0
* Created: 10/22/2009 08:40:34 PM
* Revision: none
* Compiler: gcc
*
* Author: Liang Wang (lw2aw), lw2aw@virginia.edu
* Company: CS@UVa
*
* ===... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /*
* =====================================================================================
*
* Filename: lud.cu
*
* Description: The main wrapper for the suite
*
* Version: 1.0
* Created: 10/22/2009 08:40:34 PM
* Revision: none
* Compiler: gcc
*
* Author: Liang Wang (lw2aw), lw2aw@virginia.edu
* Company: CS@UVa
*
* ===... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12lud_diagonalPfii
.globl _Z12lud_diagonalPfii
.p2align 8
.type _Z12lud_diagonalPfii,@function
_Z12lud_diagonalPfii:
s_load_b128 s[4:7], s[0:1], 0x0
v_lshlrev_b32_e32 v3, 2, v0
s_waitcnt lgkmcnt(0)
s_add_i32 s0, s6, 1
s_delay_alu inst... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include<stdio.h>
#include"cuda_runtime.h"
#include"device_launch_parameters.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
... | code for sm_80
Function : _Z3addPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_TID.X ; /* 0x0000000000067919 */
/* 0x000e220000002100 */
/*0020*/ HFMA... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include<stdio.h>
#include"cuda_runtime.h"
#include"device_launch_parameters.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
... | .file "tmpxft_000e9f40_00000000-6_w61.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include<stdio.h>
#include"cuda_runtime.h"
#include"device_launch_parameters.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
... | #include<stdio.h>
#include"hip/hip_runtime.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
{
scanf("%d",&a[i]);
}
printf("\n ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include<stdio.h>
#include"hip/hip_runtime.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
{
scanf("%d",&a[i]);
}
printf("\n ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPiS_S_
.globl _Z3addPiS_S_
.p2align 8
.type _Z3addPiS_S_,@function
_Z3addPiS_S_:
s_load_b128 s[4:7], s[0:1], 0x0
v_lshlrev_b32_e32 v0, 2, v0
s_load_b64 s[0:1], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_clause 0x1
global_... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include<stdio.h>
#include"hip/hip_runtime.h"
__global__ void add(int *a,int *b,int *c)
{
int tid=threadIdx.x;
c[tid]=a[tid]+b[tid];
}
int main()
{
int n,a[10],b[10],c[10];
printf("\nValue of N:");
scanf("%d",&n);
printf("\n Enter the Values of array A:");
for (int i = 0; i < n; ++i)
{
scanf("%d",&a[i]);
}
printf("\n ... | .text
.file "w61.hip"
.globl _Z18__device_stub__addPiS_S_ # -- Begin function _Z18__device_stub__addPiS_S_
.p2align 4, 0x90
.type _Z18__device_stub__addPiS_S_,@function
_Z18__device_stub__addPiS_S_: # @_Z18__device_stub__addPiS_S_
.cfi_startproc
# %bb.0:
subq $104, %rsp
.cfi_def_cfa_offset 112
movq %rdi, 7... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z3addPiS_S_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_TID.X ; /* 0x0000000000067919 */
/* 0x000e220000002100 */
/*0020*/ HFMA... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addPiS_S_
.globl _Z3addPiS_S_
.p2align 8
.type _Z3addPiS_S_,@function
_Z3addPiS_S_:
s_load_b128 s[4:7], s[0:1], 0x0
v_lshlrev_b32_e32 v0, 2, v0
s_load_b64 s[0:1], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_clause 0x1
global_... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000e9f40_00000000-6_w61.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... | .text
.file "w61.hip"
.globl _Z18__device_stub__addPiS_S_ # -- Begin function _Z18__device_stub__addPiS_S_
.p2align 4, 0x90
.type _Z18__device_stub__addPiS_S_,@function
_Z18__device_stub__addPiS_S_: # @_Z18__device_stub__addPiS_S_
.cfi_startproc
# %bb.0:
subq $104, %rsp
.cfi_def_cfa_offset 112
movq %rdi, 7... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /******************************************************************************
* PROGRAM: copyStruture
* PURPOSE: This program is a test which test the ability to transfer multilevel
* C++ structured data from host to device, modify them and transfer back.
*
*
* NAME: Vuong Pham-Duy.
* College student.
* Faculty of C... | code for sm_80
Function : _Z4funcPi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ IMAD.MOV.U32 R12, RZ, RZ, c[0x0][0x160] ; /* 0x00005800ff0c7624 */
/... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /******************************************************************************
* PROGRAM: copyStruture
* PURPOSE: This program is a test which test the ability to transfer multilevel
* C++ structured data from host to device, modify them and transfer back.
*
*
* NAME: Vuong Pham-Duy.
* College student.
* Faculty of C... | .file "tmpxft_00137eef_00000000-6_copyStructure.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /******************************************************************************
* PROGRAM: copyStruture
* PURPOSE: This program is a test which test the ability to transfer multilevel
* C++ structured data from host to device, modify them and transfer back.
*
*
* NAME: Vuong Pham-Duy.
* College student.
* Faculty of C... | /******************************************************************************
* PROGRAM: copyStruture
* PURPOSE: This program is a test which test the ability to transfer multilevel
* C++ structured data from host to device, modify them and transfer back.
*
*
* NAME: Vuong Pham-Duy.
* College student.
* Faculty of C... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /******************************************************************************
* PROGRAM: copyStruture
* PURPOSE: This program is a test which test the ability to transfer multilevel
* C++ structured data from host to device, modify them and transfer back.
*
*
* NAME: Vuong Pham-Duy.
* College student.
* Faculty of C... | .text
.file "copyStructure.hip"
.globl _Z9gpuAssert10hipError_tPKcib # -- Begin function _Z9gpuAssert10hipError_tPKcib
.p2align 4, 0x90
.type _Z9gpuAssert10hipError_tPKcib,@function
_Z9gpuAssert10hipError_tPKcib: # @_Z9gpuAssert10hipError_tPKcib
.cfi_startproc
# %bb.0:
pushq %rbp
.cfi_def_cfa_offset 16
pushq... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00137eef_00000000-6_copyStructure.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... | .text
.file "copyStructure.hip"
.globl _Z9gpuAssert10hipError_tPKcib # -- Begin function _Z9gpuAssert10hipError_tPKcib
.p2align 4, 0x90
.type _Z9gpuAssert10hipError_tPKcib,@function
_Z9gpuAssert10hipError_tPKcib: # @_Z9gpuAssert10hipError_tPKcib
.cfi_startproc
# %bb.0:
pushq %rbp
.cfi_def_cfa_offset 16
pushq... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
... | code for sm_80
Function : _Z12fill_double8iP7double8
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R3, SR_CTAID.X ; /* 0x0000000000037919 */
/* 0x... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
... | .file "tmpxft_000de458_00000000-6_part2.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnre... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
... | #include <hip/hip_runtime.h>
#include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11cpy_double8iP7double8S0_
.globl _Z11cpy_double8iP7double8S0_
.p2align 8
.type _Z11cpy_double8iP7double8S0_,@function
_Z11cpy_double8iP7double8S0_:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x24
s_load_b32 s8, s[0:1], 0x0
s_add_u32 s2, s0, 24
s_addc_... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <iostream>
#include <math.h>
typedef struct
{
double x;
double y;
double z;
double w;
double xx;
double yy;
double zz;
double ww;
} double8;
// Kernel function to add the elements of two arrays
__global__
void cpy_double8(int n, double8 *src, double8 *dst)
{
int index = blockIdx.x... | .text
.file "part2.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z26__device_stub__cpy_double8iP7double8S0_ # -- Begin function _Z26__device_stub__cpy_double8iP7doub... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000de458_00000000-6_part2.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnre... | .text
.file "part2.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z26__device_stub__cpy_double8iP7double8S0_ # -- Begin function _Z26__device_stub__cpy_double8iP7doub... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <cstdio>
#include <string>
#include<cuda_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*length ... | code for sm_80
Function : _Z13matrixMulCUDAPfS_S_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220000002500 *... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <cstdio>
#include <string>
#include<cuda_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*length ... | .file "tmpxft_000c2e1a_00000000-6_matMul.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3428:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <cstdio>
#include <string>
#include<cuda_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*length ... | #include <cstdio>
#include <string>
#include<hip/hip_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*leng... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <cstdio>
#include <string>
#include<hip/hip_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*leng... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z13matrixMulCUDAPfS_S_ii
.globl _Z13matrixMulCUDAPfS_S_ii
.p2align 8
.type _Z13matrixMulCUDAPfS_S_ii,@function
_Z13matrixMulCUDAPfS_S_ii:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b64 s[2:3], s[0:1], 0x18
v_and_b32_e32 v2, 0x3ff, v0
v_bf... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <cstdio>
#include <string>
#include<hip/hip_runtime.h>
#include<sys/time.h>
#include<vector>
using namespace std;
void print_data(float* arr, int width, int length, string flag) {
int count=8;
printf("%s :\n", flag.c_str());
for (int i=0; i<count; i++) {
for(int j=0; j<count; j++) {
printf("%.3lf ", arr[i*leng... | .text
.file "matMul.hip"
.globl _Z10print_dataPfiiNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEE # -- Begin function _Z10print_dataPfiiNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEE
.p2align 4, 0x90
.type _Z10print_dataPfiiNSt7__cxx1112basic_stringIcSt11char_traitsIcESaIcEEE,@function
_Z10print_dataPfiiNSt7... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z13matrixMulCUDAPfS_S_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220000002500 *... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z13matrixMulCUDAPfS_S_ii
.globl _Z13matrixMulCUDAPfS_S_ii
.p2align 8
.type _Z13matrixMulCUDAPfS_S_ii,@function
_Z13matrixMulCUDAPfS_S_ii:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b64 s[2:3], s[0:1], 0x18
v_and_b32_e32 v2, 0x3ff, v0
v_bf... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void Somatorio( float *input, float *results, long int n ) {
extern __shared__ float sdata[];
int idx = blockIdx.x * blockDim.x + threadIdx.x, tx = threadIdx.x;
float x = 0.;
if( idx < n ) {
x = input[ idx ];
}
sdata[ tx ] = x;
__syncthreads( );
for( int offset = blockDim.x / 2; offset ... | code for sm_80
Function : _Z9SomatorioPfS_l
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e22000... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.