system stringclasses 7
values | input stringlengths 10 94.5k | output stringlengths 14 55.5k |
|---|---|---|
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00193210_00000000-6_lab5_2_1.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp... | .text
.file "lab5_2_1.hip"
.globl _Z27__device_stub__vectorsum2_1Pii # -- Begin function _Z27__device_stub__vectorsum2_1Pii
.p2align 4, 0x90
.type _Z27__device_stub__vectorsum2_1Pii,@function
_Z27__device_stub__vectorsum2_1Pii: # @_Z27__device_stub__vectorsum2_1Pii
.cfi_startproc
# %bb.0:
subq $88, %rsp
.cfi_def_cf... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
#include <stdlib.h>
#include <vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = imin + (imax - imin)/2;
if (input < v... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include <stdlib.h>
#include <vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = imin + (imax - imin)/2;
if (input < v... | .file "tmpxft_001a82c2_00000000-6_binarySearch.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#include <stdlib.h>
#include <vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = imin + (imax - imin)/2;
if (input < v... | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <hip/hip_vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <hip/hip_vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = ... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <hip/hip_vector_types.h>
#include <stdint.h>
#include <assert.h>
__device__ __host__ inline
uint32_t binarySearch(uint64_t *values, uint64_t input, uint32_t len) {
int32_t imin = 0;
int32_t imax = len-1;
while (imin <= imax) {
uint32_t imid = ... | .text
.file "binarySearch.hip"
.globl main # -- Begin function main
.p2align 4, 0x90
.type main,@function
main: # @main
.cfi_startproc
# %bb.0: # %.lr.ph.i.preheader
pushq %r14
.cfi_def_cfa_offset 16
pushq %rbx
.cfi_def_cfa_offs... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001a82c2_00000000-6_binarySearch.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, ... | .text
.file "binarySearch.hip"
.globl main # -- Begin function main
.p2align 4, 0x90
.type main,@function
main: # @main
.cfi_startproc
# %bb.0: # %.lr.ph.i.preheader
pushq %r14
.cfi_def_cfa_offset 16
pushq %rbx
.cfi_def_cfa_offs... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for each array
size_t memSize = si... | code for sm_80
Function : _Z3addiPiS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R2, SR_CTAID.X ; /* 0x0000000000027919 */
/* 0x000e280000002500 */
/*0020*/ S2... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for each array
size_t memSize = si... | .file "tmpxft_00191db2_00000000-6_saxpy_basic.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cu... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for each array
size_t memSize = si... | #include <hip/hip_runtime.h>
#include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for e... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for e... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addiPiS_i
.globl _Z3addiPiS_i
.p2align 8
.type _Z3addiPiS_i,@function
_Z3addiPiS_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b32 s3, s[0:1], 0x0
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu instid0... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <iostream>
// Kernel function to add the elements of two arrays
__global__ void add(int n, int *x, int *y, int a)
{
int tid = blockIdx.x*blockDim.x + threadIdx.x;
if(tid<n) x[tid] = a*x[tid] + y[tid];
}
int main(void)
{
int dNum = 1<<24;
int *x, *y, *d_x, *d_y;
// memory size for e... | .text
.file "saxpy_basic.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z18__device_stub__addiPiS_i # -- Begin function _Z18__device_stub__addiPiS_i
.p2align 4, 0x... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z3addiPiS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R2, SR_CTAID.X ; /* 0x0000000000027919 */
/* 0x000e280000002500 */
/*0020*/ S2... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3addiPiS_i
.globl _Z3addiPiS_i
.p2align 8
.type _Z3addiPiS_i,@function
_Z3addiPiS_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b32 s3, s[0:1], 0x0
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu instid0... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00191db2_00000000-6_saxpy_basic.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cu... | .text
.file "saxpy_basic.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z18__device_stub__addiPiS_i # -- Begin function _Z18__device_stub__addiPiS_i
.p2align 4, 0x... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | __global__ void
convolution(float *N , float *M, float *P , int Tile_Size, int Mask_Width , int Width)
{
int k = blockIdx.x * blockDim.x + threadIdx.x;
int l = blockIdx.y * blockDim.y + threadIdx.y;
__shared__ float N_ds[7][7];
int n = Mask_Width/2;
int halo_index_left = (blockIdx.x - 1) * blockDim.x + threadIdx.x;
in... | .file "tmpxft_0010cdeb_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | __global__ void
convolution(float *N , float *M, float *P , int Tile_Size, int Mask_Width , int Width)
{
int k = blockIdx.x * blockDim.x + threadIdx.x;
int l = blockIdx.y * blockDim.y + threadIdx.y;
__shared__ float N_ds[7][7];
int n = Mask_Width/2;
int halo_index_left = (blockIdx.x - 1) * blockDim.x + threadIdx.x;
in... | #include <hip/hip_runtime.h>
__global__ void
convolution(float *N , float *M, float *P , int Tile_Size, int Mask_Width , int Width)
{
int k = blockIdx.x * blockDim.x + threadIdx.x;
int l = blockIdx.y * blockDim.y + threadIdx.y;
__shared__ float N_ds[7][7];
int n = Mask_Width/2;
int halo_index_left = (blockIdx.x - 1) *... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
__global__ void
convolution(float *N , float *M, float *P , int Tile_Size, int Mask_Width , int Width)
{
int k = blockIdx.x * blockDim.x + threadIdx.x;
int l = blockIdx.y * blockDim.y + threadIdx.y;
__shared__ float N_ds[7][7];
int n = Mask_Width/2;
int halo_index_left = (blockIdx.x - 1) *... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11convolutionPfS_S_iii
.globl _Z11convolutionPfS_S_iii
.p2align 8
.type _Z11convolutionPfS_S_iii,@function
_Z11convolutionPfS_S_iii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x34
s_load_b64 s[8:9], s[0:1], 0x1c
v_bfe_u32 v2, v0, 10, 10
s_load_b64... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
__global__ void
convolution(float *N , float *M, float *P , int Tile_Size, int Mask_Width , int Width)
{
int k = blockIdx.x * blockDim.x + threadIdx.x;
int l = blockIdx.y * blockDim.y + threadIdx.y;
__shared__ float N_ds[7][7];
int n = Mask_Width/2;
int halo_index_left = (blockIdx.x - 1) *... | .text
.file "kernel.hip"
.globl _Z26__device_stub__convolutionPfS_S_iii # -- Begin function _Z26__device_stub__convolutionPfS_S_iii
.p2align 4, 0x90
.type _Z26__device_stub__convolutionPfS_S_iii,@function
_Z26__device_stub__convolutionPfS_S_iii: # @_Z26__device_stub__convolutionPfS_S_iii
.cfi_startproc
# %bb.0:
subq $1... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0010cdeb_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "kernel.hip"
.globl _Z26__device_stub__convolutionPfS_S_iii # -- Begin function _Z26__device_stub__convolutionPfS_S_iii
.p2align 4, 0x90
.type _Z26__device_stub__convolutionPfS_S_iii,@function
_Z26__device_stub__convolutionPfS_S_iii: # @_Z26__device_stub__convolutionPfS_S_iii
.cfi_startproc
# %bb.0:
subq $1... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <chrono>
#include <fstream>
#include <stdio.h>
#include <string>
#include <iostream>
#include <math.h>
__constant__ int DbMToWattConstDivisor = 1000;
__constant__ double CTen = 10;
__constant__ int P = 1;
__device__
void d_dbm_to_watts( const int& dbm, double& results )
{
const double pow = (P * dbm) / CTen;
c... | .file "tmpxft_0015ac81_00000000-6_assignment.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3911:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cud... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <chrono>
#include <fstream>
#include <stdio.h>
#include <string>
#include <iostream>
#include <math.h>
__constant__ int DbMToWattConstDivisor = 1000;
__constant__ double CTen = 10;
__constant__ int P = 1;
__device__
void d_dbm_to_watts( const int& dbm, double& results )
{
const double pow = (P * dbm) / CTen;
c... | #include <hip/hip_runtime.h>
#include <chrono>
#include <fstream>
#include <stdio.h>
#include <string>
#include <iostream>
#include <math.h>
__constant__ int DbMToWattConstDivisor = 1000;
__constant__ double CTen = 10;
__constant__ int P = 1;
__device__
void d_dbm_to_watts( const int& dbm, double& results )
{
const dou... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <chrono>
#include <fstream>
#include <stdio.h>
#include <string>
#include <iostream>
#include <math.h>
__constant__ int DbMToWattConstDivisor = 1000;
__constant__ double CTen = 10;
__constant__ int P = 1;
__device__
void d_dbm_to_watts( const int& dbm, double& results )
{
const dou... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z20convert_dbm_to_wattsPKiPd
.globl _Z20convert_dbm_to_wattsPKiPd
.p2align 8
.type _Z20convert_dbm_to_wattsPKiPd,@function
_Z20convert_dbm_to_wattsPKiPd:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x1c
s_load_b128 s[4:7], s[0:1], 0x0
s_mov_b32 s3, 0x3fba... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <chrono>
#include <fstream>
#include <stdio.h>
#include <string>
#include <iostream>
#include <math.h>
__constant__ int DbMToWattConstDivisor = 1000;
__constant__ double CTen = 10;
__constant__ int P = 1;
__device__
void d_dbm_to_watts( const int& dbm, double& results )
{
const dou... | .text
.file "assignment.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z35__device_stub__convert_dbm_to_wattsPKiPd # -- Begin function _Z35__device_stub__convert_dbm_... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float shared_data[BLOCK_SIZE];
shared... | code for sm_80
Function : _Z15reductionKernelPfS_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e220000002500 */
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float shared_data[BLOCK_SIZE];
shared... | .file "tmpxft_00046f88_00000000-6_reduction.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float shared_data[BLOCK_SIZE];
shared... | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float sh... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float sh... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z15reductionKernelPfS_
.globl _Z15reductionKernelPfS_
.p2align 8
.type _Z15reductionKernelPfS_,@function
_Z15reductionKernelPfS_:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x1c
s_load_b64 s[4:5], s[0:1], 0x0
s_mov_b32 s2, s15
s_delay_alu instid0(S... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #define BLOCK_SIZE 32
#define DATALEN_PER_BLOCK (BLOCK_SIZE * 2)
#include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void reductionKernel(float *a, float *r)
{
int blocksz = blockDim.x;
int tid = threadIdx.x;
int bid = blockIdx.x;
int i1 = DATALEN_PER_BLOCK * bid + tid;
int i2 = i1 + blocksz;
__shared__ float sh... | .text
.file "reduction.hip"
.globl _Z30__device_stub__reductionKernelPfS_ # -- Begin function _Z30__device_stub__reductionKernelPfS_
.p2align 4, 0x90
.type _Z30__device_stub__reductionKernelPfS_,@function
_Z30__device_stub__reductionKernelPfS_: # @_Z30__device_stub__reductionKernelPfS_
.cfi_startproc
# %bb.0:
subq $88,... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z15reductionKernelPfS_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e220000002500 */
... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z15reductionKernelPfS_
.globl _Z15reductionKernelPfS_
.p2align 8
.type _Z15reductionKernelPfS_,@function
_Z15reductionKernelPfS_:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x1c
s_load_b64 s[4:5], s[0:1], 0x0
s_mov_b32 s2, s15
s_delay_alu instid0(S... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00046f88_00000000-6_reduction.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "reduction.hip"
.globl _Z30__device_stub__reductionKernelPfS_ # -- Begin function _Z30__device_stub__reductionKernelPfS_
.p2align 4, 0x90
.type _Z30__device_stub__reductionKernelPfS_,@function
_Z30__device_stub__reductionKernelPfS_: # @_Z30__device_stub__reductionKernelPfS_
.cfi_startproc
# %bb.0:
subq $88,... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | //pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic subtraction (final should be... | code for sm_80
Function : _Z10testKernelPi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R3, SR_CTAID.X ; /* 0x0000000000037919 */
/* 0x000e220000... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | //pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic subtraction (final should be... | .file "tmpxft_0003614f_00000000-6_simpleAtomicIntrinsics.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | //pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic subtraction (final should be... | #include <hip/hip_runtime.h>
//pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
//pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10testKernelPi
.globl _Z10testKernelPi
.p2align 8
.type _Z10testKernelPi,@function
_Z10testKernelPi:
s_clause 0x1
s_load_b64 s[2:3], s[0:1], 0x0
s_load_b32 s0, s[0:1], 0x14
s_mov_b32 s4, exec_lo
s_mov_b32 s1, exec_lo
v_mbcnt_l... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
//pass
//--gridDim=64 --blockDim=256
__global__ void
testKernel(int *g_odata)
{
// access thread id
const unsigned int tid = blockDim.x * blockIdx.x + threadIdx.x;
// Test various atomic instructions
// Arithmetic atomic instructions
// Atomic addition
atomicAdd(&g_odata[0], 10);
// Atomic... | .text
.file "simpleAtomicIntrinsics.hip"
.globl _Z25__device_stub__testKernelPi # -- Begin function _Z25__device_stub__testKernelPi
.p2align 4, 0x90
.type _Z25__device_stub__testKernelPi,@function
_Z25__device_stub__testKernelPi: # @_Z25__device_stub__testKernelPi
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z10testKernelPi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R3, SR_CTAID.X ; /* 0x0000000000037919 */
/* 0x000e220000... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10testKernelPi
.globl _Z10testKernelPi
.p2align 8
.type _Z10testKernelPi,@function
_Z10testKernelPi:
s_clause 0x1
s_load_b64 s[2:3], s[0:1], 0x0
s_load_b32 s0, s[0:1], 0x14
s_mov_b32 s4, exec_lo
s_mov_b32 s1, exec_lo
v_mbcnt_l... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0003614f_00000000-6_simpleAtomicIntrinsics.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... | .text
.file "simpleAtomicIntrinsics.hip"
.globl _Z25__device_stub__testKernelPi # -- Begin function _Z25__device_stub__testKernelPi
.p2align 4, 0x90
.type _Z25__device_stub__testKernelPi,@function
_Z25__device_stub__testKernelPi: # @_Z25__device_stub__testKernelPi
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <cuda.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * blockDim.x + thr... | code for sm_80
Function : _Z11_slowKernelPci
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2800... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <cuda.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * blockDim.x + thr... | .file "tmpxft_000ce073_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <cuda.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * blockDim.x + thr... | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <hip/hip_runtime.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * block... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <hip/hip_runtime.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * block... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11_slowKernelPci
.globl _Z11_slowKernelPci
.p2align 8
.type _Z11_slowKernelPci,@function
_Z11_slowKernelPci:
s_clause 0x1
s_load_b32 s5, s[0:1], 0x1c
s_load_b32 s4, s[0:1], 0x8
s_add_u32 s2, s0, 16
s_addc_u32 s3, s1, 0
s_waitcnt l... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /*
* Copyright (c) Meta Platforms, Inc. and affiliates.
* All rights reserved.
*
* This source code is licensed under the BSD-style license found in the
* LICENSE file in the root directory of this source tree.
*/
#include <hip/hip_runtime.h>
__global__ void _slowKernel(char* ptr, int sz) {
int idx = blockIdx.x * block... | .text
.file "kernel.hip"
.globl _Z26__device_stub___slowKernelPci # -- Begin function _Z26__device_stub___slowKernelPci
.p2align 4, 0x90
.type _Z26__device_stub___slowKernelPci,@function
_Z26__device_stub___slowKernelPci: # @_Z26__device_stub___slowKernelPci
.cfi_startproc
# %bb.0:
subq $88, %rsp
.cfi_def_cfa_offs... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z11_slowKernelPci
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2800... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11_slowKernelPci
.globl _Z11_slowKernelPci
.p2align 8
.type _Z11_slowKernelPci,@function
_Z11_slowKernelPci:
s_clause 0x1
s_load_b32 s5, s[0:1], 0x1c
s_load_b32 s4, s[0:1], 0x8
s_add_u32 s2, s0, 16
s_addc_u32 s3, s1, 0
s_waitcnt l... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000ce073_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "kernel.hip"
.globl _Z26__device_stub___slowKernelPci # -- Begin function _Z26__device_stub___slowKernelPci
.p2align 4, 0x90
.type _Z26__device_stub___slowKernelPci,@function
_Z26__device_stub___slowKernelPci: # @_Z26__device_stub___slowKernelPci
.cfi_startproc
# %bb.0:
subq $88, %rsp
.cfi_def_cfa_offs... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | __global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + threadIdx.x;
j = blockIdx.y * b... | code for sm_80
Function : _Z6mandeliiPii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R13, SR_CTAID.Y ; /* 0x00000000000d7919 */
/* 0x000e2800000... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | __global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + threadIdx.x;
j = blockIdx.y * b... | .file "tmpxft_000f7353_00000000-6_mandelgpu.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | __global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + threadIdx.x;
j = blockIdx.y * b... | #include <hip/hip_runtime.h>
__global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + th... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
__global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + th... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6mandeliiPii
.globl _Z6mandeliiPii
.p2align 8
.type _Z6mandeliiPii,@function
_Z6mandeliiPii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x0
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 v3, v0, 10, 10
s_wa... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
__global__
void mandel(int disp_width, int disp_height, int *array, int max_iter)
{
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
// Get thread indexes
i = blockIdx.x * blockDim.x + th... | .text
.file "mandelgpu.hip"
.globl _Z21__device_stub__mandeliiPii # -- Begin function _Z21__device_stub__mandeliiPii
.p2align 4, 0x90
.type _Z21__device_stub__mandeliiPii,@function
_Z21__device_stub__mandeliiPii: # @_Z21__device_stub__mandeliiPii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset 128
... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6mandeliiPii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R13, SR_CTAID.Y ; /* 0x00000000000d7919 */
/* 0x000e2800000... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6mandeliiPii
.globl _Z6mandeliiPii
.p2align 8
.type _Z6mandeliiPii,@function
_Z6mandeliiPii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x0
v_and_b32_e32 v2, 0x3ff, v0
v_bfe_u32 v3, v0, 10, 10
s_wa... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000f7353_00000000-6_mandelgpu.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "mandelgpu.hip"
.globl _Z21__device_stub__mandeliiPii # -- Begin function _Z21__device_stub__mandeliiPii
.p2align 4, 0x90
.type _Z21__device_stub__mandeliiPii,@function
_Z21__device_stub__mandeliiPii: # @_Z21__device_stub__mandeliiPii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset 128
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
gettimeofday(&ret, NULL);
return... | code for sm_80
Function : _Z20MatrixMultiplicationPfS_S_iiiiii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
gettimeofday(&ret, NULL);
return... | .file "tmpxft_001b0e72_00000000-6_MatrixMultTiled_v4.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2068:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
gettimeofday(&ret, NULL);
return... | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
get... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
get... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z20MatrixMultiplicationPfS_S_iiiiii
.globl _Z20MatrixMultiplicationPfS_S_iiiiii
.p2align 8
.type _Z20MatrixMultiplicationPfS_S_iiiiii,@function
_Z20MatrixMultiplicationPfS_S_iiiiii:
s_load_b32 s3, s[0:1], 0x1c
v_bfe_u32 v1, v0, 10, 10
v_and_b32_e32 v0, 0... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | //source: https://github.com/lzhengchun/matrix-cuda/blob/master/matrix_cuda.cu
#include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <sys/time.h>
#include <assert.h>
#define TYPE float
#define TILE_DIM 32
/*
Returns the current time in miliseconds.
*/
double getMilitime(){
struct timeval ret;
get... | .text
.file "MatrixMultTiled_v4.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function _Z11getMilitimev
.LCPI0_0:
.quad 0x412e848000000000 # double 1.0E+6
.text
.globl _Z11getMilitimev
.p2align 4, 0x90
.type _Z11getMilitimev,@function
_Z11getMilitimev: ... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z20MatrixMultiplicationPfS_S_iiiiii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z20MatrixMultiplicationPfS_S_iiiiii
.globl _Z20MatrixMultiplicationPfS_S_iiiiii
.p2align 8
.type _Z20MatrixMultiplicationPfS_S_iiiiii,@function
_Z20MatrixMultiplicationPfS_S_iiiiii:
s_load_b32 s3, s[0:1], 0x1c
v_bfe_u32 v1, v0, 10, 10
v_and_b32_e32 v0, 0... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001b0e72_00000000-6_MatrixMultTiled_v4.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2068:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... | .text
.file "MatrixMultTiled_v4.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function _Z11getMilitimev
.LCPI0_0:
.quad 0x412e848000000000 # double 1.0E+6
.text
.globl _Z11getMilitimev
.p2align 4, 0x90
.type _Z11getMilitimev,@function
_Z11getMilitimev: ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "cuda_runtime.h"
#include "device_launch_parameters.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("... | code for sm_80
Function : _Z6togglePc
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_TID.X ; /* 0x0000000000007919 */
/* 0x000e220000002100 ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "cuda_runtime.h"
#include "device_launch_parameters.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("... | .file "tmpxft_001b64d4_00000000-6_arrToggle.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "cuda_runtime.h"
#include "device_launch_parameters.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("... | #include "hip/hip_runtime.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("Enter the string to be toggled:\n")... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include "hip/hip_runtime.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("Enter the string to be toggled:\n")... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6togglePc
.globl _Z6togglePc
.p2align 8
.type _Z6togglePc,@function
_Z6togglePc:
s_load_b64 s[0:1], s[0:1], 0x0
s_mov_b32 s2, 0
s_waitcnt lgkmcnt(0)
global_load_u8 v2, v0, s[0:1]
s_waitcnt vmcnt(0)
v_add_nc_u16 v1, v... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include "hip/hip_runtime.h"
#include <stdio.h>
#include <string.h>
__global__ void toggle(char *a){
int tid = threadIdx.x;
if (a[tid] >= 'A' && a[tid] <= 'Z')
{a[tid]+=32;}
else if(a[tid] >= 'a' && a[tid] <= 'z')
{a[tid] -=32;}
}
int main(void){
int n,size;
char a[100],*d_a;
printf("Enter the string to be toggled:\n")... | .text
.file "arrToggle.hip"
.globl _Z21__device_stub__togglePc # -- Begin function _Z21__device_stub__togglePc
.p2align 4, 0x90
.type _Z21__device_stub__togglePc,@function
_Z21__device_stub__togglePc: # @_Z21__device_stub__togglePc
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_cfa_offset 80
movq %rdi, ... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6togglePc
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_TID.X ; /* 0x0000000000007919 */
/* 0x000e220000002100 ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6togglePc
.globl _Z6togglePc
.p2align 8
.type _Z6togglePc,@function
_Z6togglePc:
s_load_b64 s[0:1], s[0:1], 0x0
s_mov_b32 s2, 0
s_waitcnt lgkmcnt(0)
global_load_u8 v2, v0, s[0:1]
s_waitcnt vmcnt(0)
v_add_nc_u16 v1, v... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001b64d4_00000000-6_arrToggle.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "arrToggle.hip"
.globl _Z21__device_stub__togglePc # -- Begin function _Z21__device_stub__togglePc
.p2align 4, 0x90
.type _Z21__device_stub__togglePc,@function
_Z21__device_stub__togglePc: # @_Z21__device_stub__togglePc
.cfi_startproc
# %bb.0:
subq $72, %rsp
.cfi_def_cfa_offset 80
movq %rdi, ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include<iostream>
#include <cuda.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for device copies ... | code for sm_80
Function : _Z11sumThredBlkPi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R5, SR_TID.X ; /* 0x0000000000057919 */
/* 0x000e220000002100 */
/*0020*/... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include<iostream>
#include <cuda.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for device copies ... | .file "tmpxft_000f2191_00000000-6_task2.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnre... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include<iostream>
#include <cuda.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for device copies ... | #include<iostream>
#include <hip/hip_runtime.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for dev... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include<iostream>
#include <hip/hip_runtime.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for dev... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11sumThredBlkPi
.globl _Z11sumThredBlkPi
.p2align 8
.type _Z11sumThredBlkPi,@function
_Z11sumThredBlkPi:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x14
s_load_b64 s[0:1], s[0:1], 0x0
v_add_nc_u32_e32 v3, s15, v0
s_waitcnt lgkmcnt(0)
s... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include<iostream>
#include <hip/hip_runtime.h>
#define N 16
__global__ void sumThredBlk(int* dA) {
int index = threadIdx.x + blockIdx.x * blockDim.x;
dA[index] = threadIdx.x + blockIdx.x;
}
int main(void) {
int* hA; // host copies of a
int* dA; // device copies of a
int size = N * sizeof(int);
// Alloc space for dev... | .text
.file "task2.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z26__device_stub__sumThredBlkPi # -- Begin function _Z26__device_stub__sumThredBlkPi
.p2align 4, 0x9... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z11sumThredBlkPi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R5, SR_TID.X ; /* 0x0000000000057919 */
/* 0x000e220000002100 */
/*0020*/... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z11sumThredBlkPi
.globl _Z11sumThredBlkPi
.p2align 8
.type _Z11sumThredBlkPi,@function
_Z11sumThredBlkPi:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x14
s_load_b64 s[0:1], s[0:1], 0x0
v_add_nc_u32_e32 v3, s15, v0
s_waitcnt lgkmcnt(0)
s... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000f2191_00000000-6_task2.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnre... | .text
.file "task2.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z26__device_stub__sumThredBlkPi # -- Begin function _Z26__device_stub__sumThredBlkPi
.p2align 4, 0x9... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
sh_array[threadIdx.x] += d_arr... | code for sm_80
Function : _Z14reduce_momentsPfS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
sh_array[threadIdx.x] += d_arr... | .file "tmpxft_000810dc_00000000-6_reduce_moments.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
sh_array[threadIdx.x] += d_arr... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
s... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
s... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z14reduce_momentsPfS_i
.globl _Z14reduce_momentsPfS_i
.p2align 8
.type _Z14reduce_momentsPfS_i,@function
_Z14reduce_momentsPfS_i:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x24
s_load_b32 s4, s[0:1], 0x10
s_mov_b32 s2, s15
s_waitcnt lgkmcnt(0)
s_a... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce_moments(float *d_arr, float *d_results, int N)
{
__shared__ float sh_array[pThreads];
int n = blockDim.x * blockIdx.x + threadIdx.x;
// sh_array[threadIdx.x] = 0;
if (n < N){
for (int s = blockDim.x / 2; s > 0; s >>= 1){
if ( threadIdx.x < s)
{
s... | .text
.file "reduce_moments.hip"
.globl _Z29__device_stub__reduce_momentsPfS_i # -- Begin function _Z29__device_stub__reduce_momentsPfS_i
.p2align 4, 0x90
.type _Z29__device_stub__reduce_momentsPfS_i,@function
_Z29__device_stub__reduce_momentsPfS_i: # @_Z29__device_stub__reduce_momentsPfS_i
.cfi_startproc
# %bb.0:
subq... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z14reduce_momentsPfS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z14reduce_momentsPfS_i
.globl _Z14reduce_momentsPfS_i
.p2align 8
.type _Z14reduce_momentsPfS_i,@function
_Z14reduce_momentsPfS_i:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x24
s_load_b32 s4, s[0:1], 0x10
s_mov_b32 s2, s15
s_waitcnt lgkmcnt(0)
s_a... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000810dc_00000000-6_reduce_moments.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... | .text
.file "reduce_moments.hip"
.globl _Z29__device_stub__reduce_momentsPfS_i # -- Begin function _Z29__device_stub__reduce_momentsPfS_i
.p2align 4, 0x90
.type _Z29__device_stub__reduce_momentsPfS_i,@function
_Z29__device_stub__reduce_momentsPfS_i: # @_Z29__device_stub__reduce_momentsPfS_i
.cfi_startproc
# %bb.0:
subq... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | __global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w * y + w * h * z;
val = img[i]... | code for sm_80
Function : _Z4initPfS_S_S_S_S_iii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/* 0x000e220000002600 */
/... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | __global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w * y + w * h * z;
val = img[i]... | .file "tmpxft_0016d743_00000000-6_rof_init.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | __global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w * y + w * h * z;
val = img[i]... | #include <hip/hip_runtime.h>
__global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
__global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z4initPfS_S_S_S_S_iii
.globl _Z4initPfS_S_S_S_S_iii
.p2align 8
.type _Z4initPfS_S_S_S_S_iii,@function
_Z4initPfS_S_S_S_S_iii:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x4c
s_load_b64 s[12:13], s[0:1], 0x30
s_load_b32 s16, s[0:1], 0x38
v_bfe_u32... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
__global__ void init(float* xbar, float* xcur, float* xn, float* y1, float* y2, float* img, int w, int h, int nc) {
int x = threadIdx.x + blockDim.x * blockIdx.x;
int y = threadIdx.y + blockDim.y * blockIdx.y;
if (x < w && y < h) {
int i;
float val;
for (int z = 0; z < nc; z++) {
i = x + w ... | .text
.file "rof_init.hip"
.globl _Z19__device_stub__initPfS_S_S_S_S_iii # -- Begin function _Z19__device_stub__initPfS_S_S_S_S_iii
.p2align 4, 0x90
.type _Z19__device_stub__initPfS_S_S_S_S_iii,@function
_Z19__device_stub__initPfS_S_S_S_S_iii: # @_Z19__device_stub__initPfS_S_S_S_S_iii
.cfi_startproc
# %bb.0:
subq $168,... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z4initPfS_S_S_S_S_iii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/* 0x000e220000002600 */
/... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z4initPfS_S_S_S_S_iii
.globl _Z4initPfS_S_S_S_S_iii
.p2align 8
.type _Z4initPfS_S_S_S_S_iii,@function
_Z4initPfS_S_S_S_S_iii:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x4c
s_load_b64 s[12:13], s[0:1], 0x30
s_load_b32 s16, s[0:1], 0x38
v_bfe_u32... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0016d743_00000000-6_rof_init.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp... | .text
.file "rof_init.hip"
.globl _Z19__device_stub__initPfS_S_S_S_S_iii # -- Begin function _Z19__device_stub__initPfS_S_S_S_S_iii
.p2align 4, 0x90
.type _Z19__device_stub__initPfS_S_S_S_S_iii,@function
_Z19__device_stub__initPfS_S_S_S_S_iii: # @_Z19__device_stub__initPfS_S_S_S_S_iii
.cfi_startproc
# %bb.0:
subq $168,... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
cudaGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} | code for sm_80
Function : _Z8mykernelv
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ EXIT ; /* 0x000000000000794d */
/* 0x000fea0003800000 */
/*0020*/ BRA 0x20; /* 0xf... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
cudaGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} | .file "tmpxft_0004ef92_00000000-6_hello.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.c... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
cudaGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} | #include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
hipGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
hipGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z8mykernelv
.globl _Z8mykernelv
.p2align 8
.type _Z8mykernelv,@function
_Z8mykernelv:
s_endpgm
.section .rodata,"a",@progbits
.p2align 6, 0x0
.amdhsa_kernel _Z8mykernelv
.amdhsa_group_segment_fixed_size 0
.amdhsa_priv... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void mykernel(void){
}
int main(void){
int deviceCount;
mykernel<<<1,1>>>();
hipGetDeviceCount(&deviceCount);
printf("Hello World! Total Device: %d\n", deviceCount);
return 0;
} | .text
.file "hello.hip"
.globl _Z23__device_stub__mykernelv # -- Begin function _Z23__device_stub__mykernelv
.p2align 4, 0x90
.type _Z23__device_stub__mykernelv,@function
_Z23__device_stub__mykernelv: # @_Z23__device_stub__mykernelv
.cfi_startproc
# %bb.0:
subq $56, %rsp
.cfi_def_cfa_offset 64
leaq 32(%rsp... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z8mykernelv
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ EXIT ; /* 0x000000000000794d */
/* 0x000fea0003800000 */
/*0020*/ BRA 0x20; /* 0xf... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z8mykernelv
.globl _Z8mykernelv
.p2align 8
.type _Z8mykernelv,@function
_Z8mykernelv:
s_endpgm
.section .rodata,"a",@progbits
.p2align 6, 0x0
.amdhsa_kernel _Z8mykernelv
.amdhsa_group_segment_fixed_size 0
.amdhsa_priv... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0004ef92_00000000-6_hello.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.c... | .text
.file "hello.hip"
.globl _Z23__device_stub__mykernelv # -- Begin function _Z23__device_stub__mykernelv
.p2align 4, 0x90
.type _Z23__device_stub__mykernelv,@function
_Z23__device_stub__mykernelv: # @_Z23__device_stub__mykernelv
.cfi_startproc
# %bb.0:
subq $56, %rsp
.cfi_def_cfa_offset 64
leaq 32(%rsp... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
cudaError_t err = stmt; \
if (err != cudaSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", cudaGetErrorString(err)); \
return -1; \
} \
} while (0)
__global__ void vecAdd(float ... | code for sm_80
Function : _Z6vecAddPfS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e280000002500 */
/*0020*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
cudaError_t err = stmt; \
if (err != cudaSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", cudaGetErrorString(err)); \
return -1; \
} \
} while (0)
__global__ void vecAdd(float ... | .file "tmpxft_00175120_00000000-6_main.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2062:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cf... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
cudaError_t err = stmt; \
if (err != cudaSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", cudaGetErrorString(err)); \
return -1; \
} \
} while (0)
__global__ void vecAdd(float ... | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
hipError_t err = stmt; \
if (err != hipSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", hipGetErrorString(err)); \
return -1; \
} \
} while (0)
__g... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
hipError_t err = stmt; \
if (err != hipSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", hipGetErrorString(err)); \
return -1; \
} \
} while (0)
__g... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6vecAddPfS_S_i
.globl _Z6vecAddPfS_S_i
.p2align 8
.type _Z6vecAddPfS_S_i,@function
_Z6vecAddPfS_S_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b32 s3, s[0:1], 0x18
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_al... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <math.h>
#define BLOCK_SIZE 512
#define _check(stmt) \
do { \
hipError_t err = stmt; \
if (err != hipSuccess) { \
printf("Failed to run stmt ", #stmt); \
printf("Got CUDA error ... ", hipGetErrorString(err)); \
return -1; \
} \
} while (0)
__g... | .text
.file "main.hip"
.globl _Z21__device_stub__vecAddPfS_S_i # -- Begin function _Z21__device_stub__vecAddPfS_S_i
.p2align 4, 0x90
.type _Z21__device_stub__vecAddPfS_S_i,@function
_Z21__device_stub__vecAddPfS_S_i: # @_Z21__device_stub__vecAddPfS_S_i
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset 12... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6vecAddPfS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e280000002500 */
/*0020*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6vecAddPfS_S_i
.globl _Z6vecAddPfS_S_i
.p2align 8
.type _Z6vecAddPfS_S_i,@function
_Z6vecAddPfS_S_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b32 s3, s[0:1], 0x18
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_al... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.