system stringclasses 7
values | input stringlengths 10 94.5k | output stringlengths 14 55.5k |
|---|---|---|
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <hipfft/hipfft.h>
#include <stdio.h>
#include <math.h>
#include <stdlib.h>
#define FFTSIZE 8
#define BATCH 2
/********************/
/* CUDA ERROR CHECK */
/********************/
#define gpuErrchk(ans) { gpuAssert((ans), __FILE__, __LINE__); }
inline void gpuAssert(hipError_t code, ... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <hipfft/hipfft.h>
#include <stdio.h>
#include <math.h>
#include <stdlib.h>
#define FFTSIZE 8
#define BATCH 2
/********************/
/* CUDA ERROR CHECK */
/********************/
#define gpuErrchk(ans) { gpuAssert((ans), __FILE__, __LINE__); }
inline void gpuAssert(hipError_t code, ... | .text
.file "test1_planmany.hip"
.globl main # -- Begin function main
.p2align 4, 0x90
.type main,@function
main: # @main
.cfi_startproc
# %bb.0:
pushq %rbp
.cfi_def_cfa_offset 16
pushq %r15
.cfi_def_cfa_offset 24
pushq %r14
.cfi_def_cfa_offset 32
pushq %r13
... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000444c9_00000000-6_test1_planmany.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2084:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... | .text
.file "test1_planmany.hip"
.globl main # -- Begin function main
.p2align 4, 0x90
.type main,@function
main: # @main
.cfi_startproc
# %bb.0:
pushq %rbp
.cfi_def_cfa_offset 16
pushq %r15
.cfi_def_cfa_offset 24
pushq %r14
.cfi_def_cfa_offset 32
pushq %r13
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} | code for sm_80
Function : _Z12skip_res_addmPfS_S_m
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x00... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} | .file "tmpxft_0004696a_00000000-6_skip_res_add.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12skip_res_addmPfS_S_m
.globl _Z12skip_res_addmPfS_S_m
.p2align 8
.type _Z12skip_res_addmPfS_S_m,@function
_Z12skip_res_addmPfS_S_m:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x34
s_load_b64 s[2:3], s[0:1], 0x0
s_waitcnt lgkmcnt(0)
s_and_b32 s4, s... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void skip_res_add(size_t sz, float_t* f5, float* f1, float_t* skip_out_sum, size_t stride)
{
size_t index = blockDim.x * blockIdx.x + threadIdx.x;
if(index < sz)
{
skip_out_sum[index] += f5[index+stride];
f1[index] += f5[index];
}
} | .text
.file "skip_res_add.hip"
.globl _Z27__device_stub__skip_res_addmPfS_S_m # -- Begin function _Z27__device_stub__skip_res_addmPfS_S_m
.p2align 4, 0x90
.type _Z27__device_stub__skip_res_addmPfS_S_m,@function
_Z27__device_stub__skip_res_addmPfS_S_m: # @_Z27__device_stub__skip_res_addmPfS_S_m
.cfi_startproc
# %bb.0:
s... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z12skip_res_addmPfS_S_m
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x00... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12skip_res_addmPfS_S_m
.globl _Z12skip_res_addmPfS_S_m
.p2align 8
.type _Z12skip_res_addmPfS_S_m,@function
_Z12skip_res_addmPfS_S_m:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x34
s_load_b64 s[2:3], s[0:1], 0x0
s_waitcnt lgkmcnt(0)
s_and_b32 s4, s... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0004696a_00000000-6_skip_res_add.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, ... | .text
.file "skip_res_add.hip"
.globl _Z27__device_stub__skip_res_addmPfS_S_m # -- Begin function _Z27__device_stub__skip_res_addmPfS_S_m
.p2align 4, 0x90
.type _Z27__device_stub__skip_res_addmPfS_S_m,@function
_Z27__device_stub__skip_res_addmPfS_S_m: # @_Z27__device_stub__skip_res_addmPfS_S_m
.cfi_startproc
# %bb.0:
s... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y * width + x;
data_out[index]... | code for sm_80
Function : _Z18multi_scale_kernelPKfS0_Pfii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R5, SR_CTAID.Y ; /* 0x0000000000057919 */
/* 0x000e2800000... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y * width + x;
data_out[index]... | .file "tmpxft_00181614_00000000-6_multi_scale_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y * width + x;
data_out[index]... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z18multi_scale_kernelPKfS0_Pfii
.globl _Z18multi_scale_kernelPKfS0_Pfii
.p2align 8
.type _Z18multi_scale_kernelPKfS0_Pfii,@function
_Z18multi_scale_kernelPKfS0_Pfii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b64 s[4:5], s[0:1], 0x18
v_and_b32_e... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void multi_scale_kernel(const float *data_in, const float *scale, float *data_out, int width, int height) {
const int x = blockDim.x * blockIdx.x + threadIdx.x;
const int y = blockDim.y * blockIdx.y + threadIdx.y;
if (x < width && y < height) {
int index = y... | .text
.file "multi_scale_kernel.hip"
.globl _Z33__device_stub__multi_scale_kernelPKfS0_Pfii # -- Begin function _Z33__device_stub__multi_scale_kernelPKfS0_Pfii
.p2align 4, 0x90
.type _Z33__device_stub__multi_scale_kernelPKfS0_Pfii,@function
_Z33__device_stub__multi_scale_kernelPKfS0_Pfii: # @_Z33__device_stub__multi_sc... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z18multi_scale_kernelPKfS0_Pfii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R5, SR_CTAID.Y ; /* 0x0000000000057919 */
/* 0x000e2800000... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z18multi_scale_kernelPKfS0_Pfii
.globl _Z18multi_scale_kernelPKfS0_Pfii
.p2align 8
.type _Z18multi_scale_kernelPKfS0_Pfii,@function
_Z18multi_scale_kernelPKfS0_Pfii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b64 s[4:5], s[0:1], 0x18
v_and_b32_e... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00181614_00000000-6_multi_scale_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
add... | .text
.file "multi_scale_kernel.hip"
.globl _Z33__device_stub__multi_scale_kernelPKfS0_Pfii # -- Begin function _Z33__device_stub__multi_scale_kernelPKfS0_Pfii
.p2align 4, 0x90
.type _Z33__device_stub__multi_scale_kernelPKfS0_Pfii,@function
_Z33__device_stub__multi_scale_kernelPKfS0_Pfii: # @_Z33__device_stub__multi_sc... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void kernel(double *Dens, double *VradInt, double *VthetaInt, double *TemperInt, int nrad, int nsec, double *invdiffRmed, double *invdiffRsup, double *DensInt, int Adiabatic, double *Rmed, double dt, double *VradNew, double *VthetaNew, double *Energy, double *EnergyInt)
{
int j = thread... | .file "tmpxft_0012c599_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void kernel(double *Dens, double *VradInt, double *VthetaInt, double *TemperInt, int nrad, int nsec, double *invdiffRmed, double *invdiffRsup, double *DensInt, int Adiabatic, double *Rmed, double dt, double *VradNew, double *VthetaNew, double *Energy, double *EnergyInt)
{
int j = thread... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void kernel(double *Dens, double *VradInt, double *VthetaInt, double *TemperInt, int nrad, int nsec, double *invdiffRmed, double *invdiffRsup, double *DensInt, int Adiabatic, double *Rmed, double dt, double *VradNew, double *VthetaNew, double *Energy, double... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void kernel(double *Dens, double *VradInt, double *VthetaInt, double *TemperInt, int nrad, int nsec, double *invdiffRmed, double *invdiffRsup, double *DensInt, int Adiabatic, double *Rmed, double dt, double *VradNew, double *VthetaNew, double *Energy, double... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_
.globl _Z6kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_
.p2align 8
.type _Z6kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_,@function
_Z6kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_:
s_clause 0x4
s_load_b64 s[12:13], s[0:1], 0x50
s_load_b32 s2, s[0:1],... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void kernel(double *Dens, double *VradInt, double *VthetaInt, double *TemperInt, int nrad, int nsec, double *invdiffRmed, double *invdiffRsup, double *DensInt, int Adiabatic, double *Rmed, double dt, double *VradNew, double *VthetaNew, double *Energy, double... | .text
.file "kernel.hip"
.globl _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_ # -- Begin function _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_
.p2align 4, 0x90
.type _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_,@function
_Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_: # @_Z21__device_st... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0012c599_00000000-6_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "kernel.hip"
.globl _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_ # -- Begin function _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_
.p2align 4, 0x90
.type _Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_,@function
_Z21__device_stub__kernelPdS_S_S_iiS_S_S_iS_dS_S_S_S_: # @_Z21__device_st... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double4 *shacc = (double4*) shaccel... | code for sm_80
Function : _Z6reduceP7double4S0_S0_jj
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R2, SR_CTAID.X ; /* 0x0000000000027919 */
/* 0x000e220000002500 ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double4 *shacc = (double4*) shaccel... | .file "tmpxft_0015257b_00000000-6_reduce.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double4 *shacc = (double4*) shaccel... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.globl _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.p2align 8
.type _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj,@function
_Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b6... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduce(double4 *ac, double4 *ac1, double4 *ac2, unsigned int bf_real, unsigned int dimension){
unsigned int i = blockIdx.x * blockDim.x + threadIdx.x;
unsigned int k = dimension*bf_real;
double4 myacc;
extern __shared__ double4 shaccelerations[];
double... | .text
.file "reduce.hip"
.globl _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj # -- Begin function _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.p2align 4, 0x90
.type _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj,@function
_Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6reduceP7double4S0_S0_jj
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R2, SR_CTAID.X ; /* 0x0000000000027919 */
/* 0x000e220000002500 ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.globl _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.p2align 8
.type _Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj,@function
_Z6reduceP15HIP_vector_typeIdLj4EES1_S1_jj:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b6... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0015257b_00000000-6_reduce.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "reduce.hip"
.globl _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj # -- Begin function _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj
.p2align 4, 0x90
.type _Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj,@function
_Z21__device_stub__reduceP15HIP_vector_typeIdLj4EES1_S1_jj... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /// LSU EE 7700-2 (Spring 2013), GPU Microarchitecture
//
/// Homework 3
//
// Assignment in: http://www.ece.lsu.edu/koppel/gp/2013/hw03.pdf
//
/// Your Name:
#include <pthread.h>
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <ctype.h>
#include <time.h>
#incl... | .file "tmpxft_000a4712_00000000-6_hw3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2100:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /// LSU EE 7700-2 (Spring 2013), GPU Microarchitecture
//
/// Homework 3
//
// Assignment in: http://www.ece.lsu.edu/koppel/gp/2013/hw03.pdf
//
/// Your Name:
#include <pthread.h>
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <ctype.h>
#include <time.h>
#incl... | /// LSU EE 7700-2 (Spring 2013), GPU Microarchitecture
//
/// Homework 3
//
// Assignment in: http://www.ece.lsu.edu/koppel/gp/2013/hw03.pdf
//
/// Your Name:
#include <pthread.h>
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <ctype.h>
#include <time.h>
#incl... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /// LSU EE 7700-2 (Spring 2013), GPU Microarchitecture
//
/// Homework 3
//
// Assignment in: http://www.ece.lsu.edu/koppel/gp/2013/hw03.pdf
//
/// Your Name:
#include <pthread.h>
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <ctype.h>
#include <time.h>
#incl... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z17cuda_thread_startv
.globl _Z17cuda_thread_startv
.p2align 8
.type _Z17cuda_thread_startv,@function
_Z17cuda_thread_startv:
s_getpc_b64 s[4:5]
s_add_u32 s4, s4, d_app@rel32@lo+4
s_addc_u32 s5, s5, d_app@rel32@hi+12
s_load_b32 s4, s[4:5],... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /// LSU EE 7700-2 (Spring 2013), GPU Microarchitecture
//
/// Homework 3
//
// Assignment in: http://www.ece.lsu.edu/koppel/gp/2013/hw03.pdf
//
/// Your Name:
#include <pthread.h>
#include <string.h>
#include <stdio.h>
#include <stdlib.h>
#include <unistd.h>
#include <errno.h>
#include <ctype.h>
#include <time.h>
#incl... | .text
.file "hw3.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function _Z7time_fpv
.LCPI0_0:
.quad 0x3e112e0be826d695 # double 1.0000000000000001E-9
.text
.globl _Z7time_fpv
.p2align 4, 0x90
.type _Z7time_fpv,@function
_Z7time_fpv: ... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000a4712_00000000-6_hw3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2100:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi... | .text
.file "hw3.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function _Z7time_fpv
.LCPI0_0:
.quad 0x3e112e0be826d695 # double 1.0000000000000001E-9
.text
.globl _Z7time_fpv
.p2align 4, 0x90
.type _Z7time_fpv,@function
_Z7time_fpv: ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void cuConvertRGBToLABKernel(const float4* src, float4* dst, size_t stride, int width, int height, bool isNormalized)
{
const int x = blockIdx.x*blockDim.x + threadIdx.x;
const int y = blockIdx.y*blockDim.y + threadIdx.y;
int c = y*stride + x;
if (x<width && y<height)
{
// Read
float4 i... | .file "tmpxft_000c67e0_00000000-6_cuConvertRGBToLABKernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PL... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void cuConvertRGBToLABKernel(const float4* src, float4* dst, size_t stride, int width, int height, bool isNormalized)
{
const int x = blockIdx.x*blockDim.x + threadIdx.x;
const int y = blockIdx.y*blockDim.y + threadIdx.y;
int c = y*stride + x;
if (x<width && y<height)
{
// Read
float4 i... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void cuConvertRGBToLABKernel(const float4* src, float4* dst, size_t stride, int width, int height, bool isNormalized)
{
const int x = blockIdx.x*blockDim.x + threadIdx.x;
const int y = blockIdx.y*blockDim.y + threadIdx.y;
int c = y*stride + x;
if (x<width &&... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void cuConvertRGBToLABKernel(const float4* src, float4* dst, size_t stride, int width, int height, bool isNormalized)
{
const int x = blockIdx.x*blockDim.x + threadIdx.x;
const int y = blockIdx.y*blockDim.y + threadIdx.y;
int c = y*stride + x;
if (x<width &&... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z23cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib
.globl _Z23cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib
.p2align 8
.type _Z23cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib,@function
_Z23cuConvertRGBToLABKernelPK15HIP_vec... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void cuConvertRGBToLABKernel(const float4* src, float4* dst, size_t stride, int width, int height, bool isNormalized)
{
const int x = blockIdx.x*blockDim.x + threadIdx.x;
const int y = blockIdx.y*blockDim.y + threadIdx.y;
int c = y*stride + x;
if (x<width &&... | .text
.file "cuConvertRGBToLABKernel.hip"
.globl _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib # -- Begin function _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib
.p2align 4, 0x90
.type _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_mi... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000c67e0_00000000-6_cuConvertRGBToLABKernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PL... | .text
.file "cuConvertRGBToLABKernel.hip"
.globl _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib # -- Begin function _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_miib
.p2align 4, 0x90
.type _Z38__device_stub__cuConvertRGBToLABKernelPK15HIP_vector_typeIfLj4EEPS0_mi... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n;
cin>>n;
int *a= new int[n*n... | code for sm_80
Function : _Z6matMulPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.Y ; /* 0x0000000000047919 */
/* 0x000e220000002600 */
/*0020*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n;
cin>>n;
int *a= new int[n*n... | .file "tmpxft_00095824_00000000-6_A2_MATRIX_MATRIX.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n;
cin>>n;
int *a= new int[n*n... | #include <hip/hip_runtime.h>
#include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6matMulPiS_S_i
.globl _Z6matMulPiS_S_i
.p2align 8
.type _Z6matMulPiS_S_i,@function
_Z6matMulPiS_S_i:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x2c
s_load_b32 s2, s[0:1], 0x18
v_bfe_u32 v2, v0, 10, 10
v_and_b32_e32 v3, 0x3ff, v0
s_w... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <iostream>
using namespace std;
__global__ void matMul(int *a, int *b, int *c, int n){
int row=blockIdx.y*blockDim.y+threadIdx.y;
int col=blockIdx.x*blockDim.x+threadIdx.x;
int sum=0;
for(int j=0;j<n;j++)
{
sum=sum+a[row*n+j]*b[j*n+col];
}
c[n*row+col]=sum;
}
int main(){
int n... | .text
.file "A2_MATRIX_MATRIX.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z21__device_stub__matMulPiS_S_i # -- Begin function _Z21__device_stub__matMulPiS_S_i
.p2a... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6matMulPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.Y ; /* 0x0000000000047919 */
/* 0x000e220000002600 */
/*0020*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6matMulPiS_S_i
.globl _Z6matMulPiS_S_i
.p2align 8
.type _Z6matMulPiS_S_i,@function
_Z6matMulPiS_S_i:
s_clause 0x1
s_load_b32 s3, s[0:1], 0x2c
s_load_b32 s2, s[0:1], 0x18
v_bfe_u32 v2, v0, 10, 10
v_and_b32_e32 v3, 0x3ff, v0
s_w... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00095824_00000000-6_A2_MATRIX_MATRIX.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call... | .text
.file "A2_MATRIX_MATRIX.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z21__device_stub__matMulPiS_S_i # -- Begin function _Z21__device_stub__matMulPiS_S_i
.p2a... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STAB_PARAM = 0.8;
const double d... | code for sm_80
Function : _Z11copy_arraysPdS_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STAB_PARAM = 0.8;
const double d... | .file "tmpxft_001494a9_00000000-6_cuda_heat.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2062:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STAB_PARAM = 0.8;
const double d... | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <hip/hip_runtime.h>
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STA... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <hip/hip_runtime.h>
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STA... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z26update_periodic_boundariesPd
.globl _Z26update_periodic_boundariesPd
.p2align 8
.type _Z26update_periodic_boundariesPd,@function
_Z26update_periodic_boundariesPd:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x14
s_load_b64 s[2:3], s[0:1], 0x0
v_bfe_u32 v1... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /*
Compile using nvcc cuda_heat.cu
Author: Romit Maulik - romit.maulik@okstate.edu
*/
#include <hip/hip_runtime.h>
#include <stdlib.h>
#include <stdio.h>
#include <math.h>
#include <time.h>
const double PI = 3.1415926535;
const double lx = 2.0*PI, ly = 2.0*PI;
const int nx = 254, ny = 254;
const double ALPHA = 0.8, STA... | .text
.file "cuda_heat.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function main
.LCPI0_0:
.quad 0x406fc00000000000 # double 254
.LCPI0_1:
.quad 0x400921fb54411744 # double 3.1415926535000001
.LCPI0_2:
.quad 0x3f240d21199b749a # ... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z11copy_arraysPdS_
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e220... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z26update_periodic_boundariesPd
.globl _Z26update_periodic_boundariesPd
.p2align 8
.type _Z26update_periodic_boundariesPd,@function
_Z26update_periodic_boundariesPd:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x14
s_load_b64 s[2:3], s[0:1], 0x0
v_bfe_u32 v1... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001494a9_00000000-6_cuda_heat.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2062:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "cuda_heat.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function main
.LCPI0_0:
.quad 0x406fc00000000000 # double 254
.LCPI0_1:
.quad 0x400921fb54411744 # double 3.1415926535000001
.LCPI0_2:
.quad 0x3f240d21199b749a # ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && cur_dis > d_d[j] + d_g[j*n+i])... | code for sm_80
Function : _Z12bellman_fordPiS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R9, SR_CTAID.X ; /* 0x0000000000097919 */
/* 0x000e... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && cur_dis > d_d[j] + d_g[j*n+i])... | .file "tmpxft_000aa007_00000000-6_sssp.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cf... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && cur_dis > d_d[j] + d_g[j*n+i])... | #include <hip/hip_runtime.h>
#include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && c... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && c... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12bellman_fordPiS_ii
.globl _Z12bellman_fordPiS_ii
.p2align 8
.type _Z12bellman_fordPiS_ii,@function
_Z12bellman_fordPiS_ii:
s_load_b32 s2, s[0:1], 0x24
v_mov_b32_e32 v6, 0
s_waitcnt lgkmcnt(0)
s_and_b32 s4, s2, 0xffff
s_load_b64 s[2:3], ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include<stdio.h>
#include<stdlib.h>
#include<time.h>
__device__ int d_change;
__global__ void bellman_ford(int *d_g, int *d_d, int k, int n)
{
d_change = 0;
int i = blockIdx.x*blockDim.x+threadIdx.x;
int cur_dis = d_d[i];
__syncthreads();
int j;
for (j=1; j<n;j++)
{
if (d_g[j*n+i]==1 && c... | .text
.file "sssp.hip"
.globl _Z27__device_stub__bellman_fordPiS_ii # -- Begin function _Z27__device_stub__bellman_fordPiS_ii
.p2align 4, 0x90
.type _Z27__device_stub__bellman_fordPiS_ii,@function
_Z27__device_stub__bellman_fordPiS_ii: # @_Z27__device_stub__bellman_fordPiS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.c... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z12bellman_fordPiS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R9, SR_CTAID.X ; /* 0x0000000000097919 */
/* 0x000e... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z12bellman_fordPiS_ii
.globl _Z12bellman_fordPiS_ii
.p2align 8
.type _Z12bellman_fordPiS_ii,@function
_Z12bellman_fordPiS_ii:
s_load_b32 s2, s[0:1], 0x24
v_mov_b32_e32 v6, 0
s_waitcnt lgkmcnt(0)
s_and_b32 s4, s2, 0xffff
s_load_b64 s[2:3], ... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000aa007_00000000-6_sssp.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cf... | .text
.file "sssp.hip"
.globl _Z27__device_stub__bellman_fordPiS_ii # -- Begin function _Z27__device_stub__bellman_fordPiS_ii
.p2align 4, 0x90
.type _Z27__device_stub__bellman_fordPiS_ii,@function
_Z27__device_stub__bellman_fordPiS_ii: # @_Z27__device_stub__bellman_fordPiS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.c... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ascending */
if (dev_values[i... | code for sm_80
Function : _Z17bitonic_sort_stepPiii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x0... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ascending */
if (dev_values[i... | .file "tmpxft_0016b46b_00000000-6_bitonic_sort_step.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ascending */
if (dev_values[i... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z17bitonic_sort_stepPiii
.globl _Z17bitonic_sort_stepPiii
.p2align 8
.type _Z17bitonic_sort_stepPiii,@function
_Z17bitonic_sort_stepPiii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x1c
s_load_b32 s3, s[0:1], 0x8
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void bitonic_sort_step(int *dev_values, int j, int k)
{
unsigned int i, ixj; /* Sorting partners: i and ixj */
i = threadIdx.x + blockDim.x * blockIdx.x;
ixj = i^j;
/* The threads with the lowest ids sort the array. */
if ((ixj)>i) {
if ((i&k)==0) {
/* Sort ... | .text
.file "bitonic_sort_step.hip"
.globl _Z32__device_stub__bitonic_sort_stepPiii # -- Begin function _Z32__device_stub__bitonic_sort_stepPiii
.p2align 4, 0x90
.type _Z32__device_stub__bitonic_sort_stepPiii,@function
_Z32__device_stub__bitonic_sort_stepPiii: # @_Z32__device_stub__bitonic_sort_stepPiii
.cfi_startproc
... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z17bitonic_sort_stepPiii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x0... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z17bitonic_sort_stepPiii
.globl _Z17bitonic_sort_stepPiii
.p2align 8
.type _Z17bitonic_sort_stepPiii,@function
_Z17bitonic_sort_stepPiii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x1c
s_load_b32 s3, s[0:1], 0x8
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0016b46b_00000000-6_bitonic_sort_step.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq... | .text
.file "bitonic_sort_step.hip"
.globl _Z32__device_stub__bitonic_sort_stepPiii # -- Begin function _Z32__device_stub__bitonic_sort_stepPiii
.p2align 4, 0x90
.type _Z32__device_stub__bitonic_sort_stepPiii,@function
_Z32__device_stub__bitonic_sort_stepPiii: # @_Z32__device_stub__bitonic_sort_stepPiii
.cfi_startproc
... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | // This exercise is for student to learn about data sharing and synchronization between threads
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * blockDim.x + threadIdx.x;
tmp[i] =... | code for sm_80
Function : _Z10vector_macPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | // This exercise is for student to learn about data sharing and synchronization between threads
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * blockDim.x + threadIdx.x;
tmp[i] =... | .file "tmpxft_000a3e85_00000000-6_Lab5-3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | // This exercise is for student to learn about data sharing and synchronization between threads
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * blockDim.x + threadIdx.x;
tmp[i] =... | // This exercise is for student to learn about data sharing and synchronization between threads
#include <hip/hip_runtime.h>
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * block... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | // This exercise is for student to learn about data sharing and synchronization between threads
#include <hip/hip_runtime.h>
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * block... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10vector_macPiS_S_i
.globl _Z10vector_macPiS_S_i
.p2align 8
.type _Z10vector_macPiS_S_i,@function
_Z10vector_macPiS_S_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b128 s[4:7], s[0:1], 0x8
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xf... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | // This exercise is for student to learn about data sharing and synchronization between threads
#include <hip/hip_runtime.h>
#include <stdio.h>
#define N 4 // number of elements in vector
__global__
void vector_mac(int *d_c, int *d_a, int *d_b, int n){
__shared__ int tmp[N]; // shared memory
int i = blockIdx.x * block... | .text
.file "Lab5-3.hip"
.globl _Z25__device_stub__vector_macPiS_S_i # -- Begin function _Z25__device_stub__vector_macPiS_S_i
.p2align 4, 0x90
.type _Z25__device_stub__vector_macPiS_S_i,@function
_Z25__device_stub__vector_macPiS_S_i: # @_Z25__device_stub__vector_macPiS_S_i
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z10vector_macPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e2... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z10vector_macPiS_S_i
.globl _Z10vector_macPiS_S_i
.p2align 8
.type _Z10vector_macPiS_S_i,@function
_Z10vector_macPiS_S_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x2c
s_load_b128 s[4:7], s[0:1], 0x8
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xf... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000a3e85_00000000-6_Lab5-3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "Lab5-3.hip"
.globl _Z25__device_stub__vector_macPiS_S_i # -- Begin function _Z25__device_stub__vector_macPiS_S_i
.p2align 4, 0x90
.type _Z25__device_stub__vector_macPiS_S_i,@function
_Z25__device_stub__vector_macPiS_S_i: # @_Z25__device_stub__vector_macPiS_S_i
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void fm_order2_dgrad_kernel(const float* in, const float* top_grad, float* dgrad, int batch_size, int slot_num, int emb_vec_size) {
int tid = threadIdx.x;
int bid = blockIdx.x;
if (tid < emb_vec_size && bid < batch_size) {
float emb_sum = 0.0f;
int offset = bid * slot_num * emb_vec_size... | .file "tmpxft_0004e39f_00000000-6_fm_order2_dgrad_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void fm_order2_dgrad_kernel(const float* in, const float* top_grad, float* dgrad, int batch_size, int slot_num, int emb_vec_size) {
int tid = threadIdx.x;
int bid = blockIdx.x;
if (tid < emb_vec_size && bid < batch_size) {
float emb_sum = 0.0f;
int offset = bid * slot_num * emb_vec_size... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void fm_order2_dgrad_kernel(const float* in, const float* top_grad, float* dgrad, int batch_size, int slot_num, int emb_vec_size) {
int tid = threadIdx.x;
int bid = blockIdx.x;
if (tid < emb_vec_size && bid < batch_size) {
float emb_sum = 0.0f;
int offset = ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void fm_order2_dgrad_kernel(const float* in, const float* top_grad, float* dgrad, int batch_size, int slot_num, int emb_vec_size) {
int tid = threadIdx.x;
int bid = blockIdx.x;
if (tid < emb_vec_size && bid < batch_size) {
float emb_sum = 0.0f;
int offset = ... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z22fm_order2_dgrad_kernelPKfS0_Pfiii
.globl _Z22fm_order2_dgrad_kernelPKfS0_Pfiii
.p2align 8
.type _Z22fm_order2_dgrad_kernelPKfS0_Pfiii,@function
_Z22fm_order2_dgrad_kernelPKfS0_Pfiii:
s_clause 0x1
s_load_b32 s6, s[0:1], 0x20
s_load_b32 s2, s[0:1], 0x18
... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void fm_order2_dgrad_kernel(const float* in, const float* top_grad, float* dgrad, int batch_size, int slot_num, int emb_vec_size) {
int tid = threadIdx.x;
int bid = blockIdx.x;
if (tid < emb_vec_size && bid < batch_size) {
float emb_sum = 0.0f;
int offset = ... | .text
.file "fm_order2_dgrad_kernel.hip"
.globl _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii # -- Begin function _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii
.p2align 4, 0x90
.type _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii,@function
_Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii: # @_Z3... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0004e39f_00000000-6_fm_order2_dgrad_kernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... | .text
.file "fm_order2_dgrad_kernel.hip"
.globl _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii # -- Begin function _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii
.p2align 4, 0x90
.type _Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii,@function
_Z37__device_stub__fm_order2_dgrad_kernelPKfS0_Pfiii: # @_Z3... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding current row in grid
+ blockDim.x*b... | code for sm_80
Function : _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/* 0x0... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding current row in grid
+ blockDim.x*b... | .file "tmpxft_0011f3e8_00000000-6_CopyPointsCoordinatesKernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinar... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding current row in grid
+ blockDim.x*b... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding curre... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding curre... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.globl _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.p2align 8
.type _Z27CopyPointsCoordinatesKernelPfPifffS_ii,@function
_Z27CopyPointsCoordinatesKernelPfPifffS_ii:
s_clause 0x1
s_load_b64 s[2:3], s[0:1], 0x2c
s_loa... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void CopyPointsCoordinatesKernel( float *pointsCoordinates, int *activityFlag, float xNonValid, float yNonValid, float zNonValid, float *dataVertex, int dataVertexOffset, int maxCells )
{
int threadId = blockDim.x*blockIdx.y*gridDim.x //rows preceeding curre... | .text
.file "CopyPointsCoordinatesKernel.hip"
.globl _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii # -- Begin function _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii
.p2align 4, 0x90
.type _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii,@function
_Z42__device_stub__CopyPointsCoordinates... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/* 0x0... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.globl _Z27CopyPointsCoordinatesKernelPfPifffS_ii
.p2align 8
.type _Z27CopyPointsCoordinatesKernelPfPifffS_ii,@function
_Z27CopyPointsCoordinatesKernelPfPifffS_ii:
s_clause 0x1
s_load_b64 s[2:3], s[0:1], 0x2c
s_loa... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0011f3e8_00000000-6_CopyPointsCoordinatesKernel.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinar... | .text
.file "CopyPointsCoordinatesKernel.hip"
.globl _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii # -- Begin function _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii
.p2align 4, 0x90
.type _Z42__device_stub__CopyPointsCoordinatesKernelPfPifffS_ii,@function
_Z42__device_stub__CopyPointsCoordinates... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
* MAC0431 - Introducao a Programacao Paralela e Distribuida
*
* Fisica Alternativa
*
* Bruno Endo - 7990982
* Danilo Aleixo - 7972370
* Gustavo Caparica - 7991020
*
*/
#include <stdlib.h>
#include <stdio.h>
#include <time.h>
#include <assert.h>
#include <cuda.h>
#include <cuda_runtime.h>
__constant__ int WIDTH, HEIG... | .file "tmpxft_000f1887_00000000-6_main.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2065:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cf... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
* MAC0431 - Introducao a Programacao Paralela e Distribuida
*
* Fisica Alternativa
*
* Bruno Endo - 7990982
* Danilo Aleixo - 7972370
* Gustavo Caparica - 7991020
*
*/
#include <stdlib.h>
#include <stdio.h>
#include <time.h>
#include <assert.h>
#include <cuda.h>
#include <cuda_runtime.h>
__constant__ int WIDTH, HEIG... | /*
* MAC0431 - Introducao a Programacao Paralela e Distribuida
*
* Fisica Alternativa
*
* Bruno Endo - 7990982
* Danilo Aleixo - 7972370
* Gustavo Caparica - 7991020
*
*/
#include <stdlib.h>
#include <stdio.h>
#include <time.h>
#include <assert.h>
#include <hip/hip_runtime.h>
__constant__ int WIDTH, HEIGHT;
int WIDTHH,... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <math.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <time.h>
#include "max_reduction.cuh"
int* arr;
static void generate_integers(const int ARR_SIZE);
static int sequential(const int ARR_SIZE);
int main(int argc, char* argv[]) {
if (argc != 2 && argc != 3) {
fprintf(stderr, "Usage: %s ... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <math.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <time.h>
#include "max_reduction.cuh"
int* arr;
static void generate_integers(const int ARR_SIZE);
static int sequential(const int ARR_SIZE);
int main(int argc, char* argv[]) {
if (argc != 2 && argc != 3) {
fprintf(stderr, "Usage: %s ... | .file "tmpxft_00171398_00000000-6_max_reduction.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2062:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <math.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <time.h>
#include "max_reduction.cuh"
int* arr;
static void generate_integers(const int ARR_SIZE);
static int sequential(const int ARR_SIZE);
int main(int argc, char* argv[]) {
if (argc != 2 && argc != 3) {
fprintf(stderr, "Usage: %s ... | #ifndef _MAX_REDUCTION_CUH_INCLUDED_
#define _MAX_REDUCTION_CUH_INCLUDED_
#include <sys/time.h>
#define GET_TIME(now) \
{ \
struct timeval t; \
gettimeofday(&t, NULL); \
now = t.tv_sec + t.tv_usec / 1000000.0; \
}
#define DEF_ARR_SIZE 10000
#define DEF_BLOCK_SIZE 512
__host__ void reduction_divergent(const int* arr, co... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #ifndef _MAX_REDUCTION_CUH_INCLUDED_
#define _MAX_REDUCTION_CUH_INCLUDED_
#include <sys/time.h>
#define GET_TIME(now) \
{ \
struct timeval t; \
gettimeofday(&t, NULL); \
now = t.tv_sec + t.tv_usec / 1000000.0; \
}
#define DEF_ARR_SIZE 10000
#define DEF_BLOCK_SIZE 512
__host__ void reduction_divergent(const int* arr, co... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #ifndef _MAX_REDUCTION_CUH_INCLUDED_
#define _MAX_REDUCTION_CUH_INCLUDED_
#include <sys/time.h>
#define GET_TIME(now) \
{ \
struct timeval t; \
gettimeofday(&t, NULL); \
now = t.tv_sec + t.tv_usec / 1000000.0; \
}
#define DEF_ARR_SIZE 10000
#define DEF_BLOCK_SIZE 512
__host__ void reduction_divergent(const int* arr, co... | .text
.file "max_reduction.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-pr... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_00171398_00000000-6_max_reduction.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2062:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8,... | .text
.file "max_reduction.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-pr... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void sum_optimization(float* in, int inStr0, int inStr1, int inStr2, int inStr3, float* out, int outStr0, int outStr1, int outStr2, int dim, int nElementOut, int dimSize) {
int tid = blockIdx.x * blockDim.x + threadIdx.x;
int stride = gridDim.x * blockDim.x;
for (int i = tid; i < nEleme... | .file "tmpxft_00032e17_00000000-6_sum_optimization.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void sum_optimization(float* in, int inStr0, int inStr1, int inStr2, int inStr3, float* out, int outStr0, int outStr1, int outStr2, int dim, int nElementOut, int dimSize) {
int tid = blockIdx.x * blockDim.x + threadIdx.x;
int stride = gridDim.x * blockDim.x;
for (int i = tid; i < nEleme... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void sum_optimization(float* in, int inStr0, int inStr1, int inStr2, int inStr3, float* out, int outStr0, int outStr1, int outStr2, int dim, int nElementOut, int dimSize) {
int tid = blockIdx.x * blockDim.x + threadIdx.x;
int stride = gridDim.x * blockDim.x;... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void sum_optimization(float* in, int inStr0, int inStr1, int inStr2, int inStr3, float* out, int outStr0, int outStr1, int outStr2, int dim, int nElementOut, int dimSize) {
int tid = blockIdx.x * blockDim.x + threadIdx.x;
int stride = gridDim.x * blockDim.x;... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z16sum_optimizationPfiiiiS_iiiiii
.globl _Z16sum_optimizationPfiiiiS_iiiiii
.p2align 8
.type _Z16sum_optimizationPfiiiiS_iiiiii,@function
_Z16sum_optimizationPfiiiiS_iiiiii:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x44
s_load_b32 s10, s[0:1], 0x30
s_add_u3... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void sum_optimization(float* in, int inStr0, int inStr1, int inStr2, int inStr3, float* out, int outStr0, int outStr1, int outStr2, int dim, int nElementOut, int dimSize) {
int tid = blockIdx.x * blockDim.x + threadIdx.x;
int stride = gridDim.x * blockDim.x;... | .text
.file "sum_optimization.hip"
.globl _Z31__device_stub__sum_optimizationPfiiiiS_iiiiii # -- Begin function _Z31__device_stub__sum_optimizationPfiiiiS_iiiiii
.p2align 4, 0x90
.type _Z31__device_stub__sum_optimizationPfiiiiS_iiiiii,@function
_Z31__device_stub__sum_optimizationPfiiiiS_iiiiii: # @_Z31__device_stub__su... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.