system stringclasses 7
values | input stringlengths 10 94.5k | output stringlengths 14 55.5k |
|---|---|---|
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | // includes, system
#include <stdlib.h>
#include <stdio.h>
#include <string.h>
#if defined(__unix) || defined(__linux)
#include <sys/time.h>
#endif
// includes, project
#include <hip/hip_runtime.h>
#include <hipfft/hipfft.h>
extern "C" void NewCUDAFFT(float *input, int dim[3], int forward, int doComplex, int time)
{
#i... | .text
.file "NewCUDAFFT.hip"
.section .rodata.cst8,"aM",@progbits,8
.p2align 3, 0x0 # -- Begin function NewCUDAFFT
.LCPI0_0:
.quad 0x412e848000000000 # double 1.0E+6
.section .rodata.cst4,"aM",@progbits,4
.p2align 2, 0x0
.LCPI0_1:
.long 0x3f800000 # float 1
.te... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(cudaError err, const char* file, const int line)
{
#if... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(cudaError err, const char* file, const int line)
{
#if... | .file "tmpxft_001a0332_00000000-6_cuda_error_check.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(cudaError err, const char* file, const int line)
{
#if... | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(hipError_t err, const cha... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(hipError_t err, const cha... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
// Define this to turn on error checking
#define CUDA_ERROR_CHECK
#define CUDASAFECALL( err ) __cudaSafeCall( err, __FILE__, __LINE__ )
#define CUDACHECKERROR() __cudaCheckError( __FILE__, __LINE__ )
inline void __cudaSafeCall(hipError_t err, const cha... | .text
.file "cuda_error_check.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001a0332_00000000-6_cuda_error_check.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... | .text
.file "cuda_error_check.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... | code for sm_80
Function : _Z3ADDPfS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e280000002500 */
/*0020*/ S2R... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... | .file "tmpxft_000d62a8_00000000-6_b.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi_d... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3ADDPfS_i
.globl _Z3ADDPfS_i
.p2align 8
.type _Z3ADDPfS_i,@function
_Z3ADDPfS_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu instid0(... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /* *
* Copyright 1993-2012 NVIDIA Corporation. All rights reserved.
*
* Please refer to the NVIDIA end user license agreement (EULA) associated
* with this source code for terms and conditions that govern your use of
* this software. Any use, reproduction, disclosure, or distribution of
* this software and related docu... | .text
.file "b.hip"
.globl _Z18__device_stub__ADDPfS_i # -- Begin function _Z18__device_stub__ADDPfS_i
.p2align 4, 0x90
.type _Z18__device_stub__ADDPfS_i,@function
_Z18__device_stub__ADDPfS_i: # @_Z18__device_stub__ADDPfS_i
.cfi_startproc
# %bb.0:
subq $104, %rsp
.cfi_def_cfa_offset 112
movq %rdi, 72(%rs... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z3ADDPfS_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e280000002500 */
/*0020*/ S2R... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z3ADDPfS_i
.globl _Z3ADDPfS_i
.p2align 8
.type _Z3ADDPfS_i,@function
_Z3ADDPfS_i:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu instid0(... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000d62a8_00000000-6_b.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.cfi_d... | .text
.file "b.hip"
.globl _Z18__device_stub__ADDPfS_i # -- Begin function _Z18__device_stub__ADDPfS_i
.p2align 4, 0x90
.type _Z18__device_stub__ADDPfS_i,@function
_Z18__device_stub__ADDPfS_i: # @_Z18__device_stub__ADDPfS_i
.cfi_startproc
# %bb.0:
subq $104, %rsp
.cfi_def_cfa_offset 112
movq %rdi, 72(%rs... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <math.h>
__device__ __host__ float dif(float x, float y, float z, float step)
{
return (x - 2 * y + z) / step / step;
}
//TODO change step and matrix_size to vector
__global__ void potential_establish(float *prev_phi, float *next_phi, float *sum, float *step, int *matrix_size )
{
// __shared__ float sums[300];... | .file "tmpxft_001056f5_00000000-6_bm_kernel_3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <math.h>
__device__ __host__ float dif(float x, float y, float z, float step)
{
return (x - 2 * y + z) / step / step;
}
//TODO change step and matrix_size to vector
__global__ void potential_establish(float *prev_phi, float *next_phi, float *sum, float *step, int *matrix_size )
{
// __shared__ float sums[300];... | #include <hip/hip_runtime.h>
#include <math.h>
__device__ __host__ float dif(float x, float y, float z, float step)
{
return (x - 2 * y + z) / step / step;
}
//TODO change step and matrix_size to vector
__global__ void potential_establish(float *prev_phi, float *next_phi, float *sum, float *step, int *matrix_size )
{
/... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <math.h>
__device__ __host__ float dif(float x, float y, float z, float step)
{
return (x - 2 * y + z) / step / step;
}
//TODO change step and matrix_size to vector
__global__ void potential_establish(float *prev_phi, float *next_phi, float *sum, float *step, int *matrix_size )
{
/... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z19potential_establishPfS_S_S_Pi
.globl _Z19potential_establishPfS_S_S_Pi
.p2align 8
.type _Z19potential_establishPfS_S_S_Pi,@function
_Z19potential_establishPfS_S_S_Pi:
s_clause 0x1
s_load_b64 s[4:5], s[0:1], 0x20
s_load_b32 s6, s[0:1], 0x34
s_mov_b3... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <math.h>
__device__ __host__ float dif(float x, float y, float z, float step)
{
return (x - 2 * y + z) / step / step;
}
//TODO change step and matrix_size to vector
__global__ void potential_establish(float *prev_phi, float *next_phi, float *sum, float *step, int *matrix_size )
{
/... | .text
.file "bm_kernel_3.hip"
.globl _Z3difffff # -- Begin function _Z3difffff
.p2align 4, 0x90
.type _Z3difffff,@function
_Z3difffff: # @_Z3difffff
.cfi_startproc
# %bb.0:
addss %xmm1, %xmm1
subss %xmm1, %xmm0
addss %xmm2, %xmm0
divss %xmm3, %xmm0
divss %xmm3, %xmm0
ret... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001056f5_00000000-6_bm_kernel_3.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %... | .text
.file "bm_kernel_3.hip"
.globl _Z3difffff # -- Begin function _Z3difffff
.p2align 4, 0x90
.type _Z3difffff,@function
_Z3difffff: # @_Z3difffff
.cfi_startproc
# %bb.0:
addss %xmm1, %xmm1
subss %xmm1, %xmm0
addss %xmm2, %xmm0
divss %xmm3, %xmm0
divss %xmm3, %xmm0
ret... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + threadIdx.x;
if ( id < elements )... | code for sm_80
Function : _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + threadIdx.x;
if ( id < elements )... | .file "tmpxft_000bc4e2_00000000-6_calcConvolutionUpdateWeightsGPU.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatB... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + threadIdx.x;
if ( id < elements )... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + thre... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + thre... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.globl _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.p2align 8
.type _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi,@function
_Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi:
s_clause 0x2
s_load_b32 s2, s[0:1]... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void calcConvolutionUpdateWeightsGPU( float *filters, float *filter_grads, int in_size_z, int number_filters, int kernel_size, float momentum, float decay, float learning_rate, int elements )
{
int id = (blockIdx.x + blockIdx.y*gridDim.x) * blockDim.x + thre... | .text
.file "calcConvolutionUpdateWeightsGPU.hip"
.globl _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi # -- Begin function _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi
.p2align 4, 0x90
.type _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi,@function
_Z46__device_stub__calcC... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.Y ; /* 0x0000000000007919 */
/*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.globl _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi
.p2align 8
.type _Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi,@function
_Z31calcConvolutionUpdateWeightsGPUPfS_iiifffi:
s_clause 0x2
s_load_b32 s2, s[0:1]... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000bc4e2_00000000-6_calcConvolutionUpdateWeightsGPU.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatB... | .text
.file "calcConvolutionUpdateWeightsGPU.hip"
.globl _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi # -- Begin function _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi
.p2align 4, 0x90
.type _Z46__device_stub__calcConvolutionUpdateWeightsGPUPfS_iiifffi,@function
_Z46__device_stub__calcC... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | // filename: freduce.cu
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//======================================================================
#ifdef __CUDA__
__dev... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | // filename: freduce.cu
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//======================================================================
#ifdef __CUDA__
__dev... | .file "tmpxft_001672bc_00000000-6_freduce.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | // filename: freduce.cu
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//======================================================================
#ifdef __CUDA__
__dev... | // filename: freduce.cu
#include <hip/hip_runtime.h>
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//================================================================... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | // filename: freduce.cu
#include <hip/hip_runtime.h>
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//================================================================... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | // filename: freduce.cu
#include <hip/hip_runtime.h>
#include <stdint.h>
//======================================================================
#ifdef __CUDA__
__device__ __host__
#endif
void reduce_hash(uint32_t H[], uint8_t B[], uint32_t link_idx);
//================================================================... | .text
.file "freduce.hip"
.globl _Z11reduce_hashPjPhj # -- Begin function _Z11reduce_hashPjPhj
.p2align 4, 0x90
.type _Z11reduce_hashPjPhj,@function
_Z11reduce_hashPjPhj: # @_Z11reduce_hashPjPhj
.cfi_startproc
# %bb.0:
movl (%rdi), %eax
addl %edx, %eax
movzwl %ax, %ecx
imulq $165191050, %rc... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001672bc_00000000-6_freduce.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2030:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
... | .text
.file "freduce.hip"
.globl _Z11reduce_hashPjPhj # -- Begin function _Z11reduce_hashPjPhj
.p2align 4, 0x90
.type _Z11reduce_hashPjPhj,@function
_Z11reduce_hashPjPhj: # @_Z11reduce_hashPjPhj
.cfi_startproc
# %bb.0:
movl (%rdi), %eax
addl %edx, %eax
movzwl %ax, %ecx
imulq $165191050, %rc... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector addition of " << numElements <<... | code for sm_80
Function : _Z6squarePKfPfi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e280000002500 */
/*0020*/... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector addition of " << numElements <<... | .file "tmpxft_000fe2a1_00000000-6_square.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnr... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector addition of " << numElements <<... | #include <hip/hip_runtime.h>
#include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector ad... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector ad... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6squarePKfPfi
.globl _Z6squarePKfPfi
.p2align 8
.type _Z6squarePKfPfi,@function
_Z6squarePKfPfi:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu ... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <iostream>
__global__
void square(const float *A, float *B, int numElements)
{
int i = blockDim.x * blockIdx.x + threadIdx.x;
if (i < numElements)
{
B[i] = A[i] * A[i];
}
}
int main(void)
{
int numElements = 50000;
size_t size = numElements * sizeof(float);
std::cout << "[Vector ad... | .text
.file "square.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z21__device_stub__squarePKfPfi # -- Begin function _Z21__device_stub__squarePKfPfi
.p2align 4, 0x90... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6squarePKfPfi
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e280000002500 */
/*0020*/... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6squarePKfPfi
.globl _Z6squarePKfPfi
.p2align 8
.type _Z6squarePKfPfi,@function
_Z6squarePKfPfi:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_delay_alu ... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000fe2a1_00000000-6_square.cudafe1.cpp"
.text
#APP
.globl _ZSt21ios_base_library_initv
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB3672:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnr... | .text
.file "square.hip"
# Start of file scope inline assembly
.globl _ZSt21ios_base_library_initv
# End of file scope inline assembly
.globl _Z21__device_stub__squarePKfPfi # -- Begin function _Z21__device_stub__squarePKfPfi
.p2align 4, 0x90... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} | code for sm_80
Function : _Z8calcdataPdS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e22000... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} | .file "tmpxft_0001cad3_00000000-6_calcdata.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} | #include <hip/hip_runtime.h>
#include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z8calcdataPdS_ii
.globl _Z8calcdataPdS_ii
.p2align 8
.type _Z8calcdataPdS_ii,@function
_Z8calcdataPdS_ii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_de... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
#define FLOAT_N 3214212.01
__global__ void calcdata(double* d_data, double* d_mean, int M, int N)
{
int j;
int i = blockDim.x * blockIdx.x + threadIdx.x+1;
if (i<=(N+1)) {
for (j = 1; j < (M+1); j++) {
d_data[i*(M+1) + j] -= d_mean[j];
}
}
} | .text
.file "calcdata.hip"
.globl _Z23__device_stub__calcdataPdS_ii # -- Begin function _Z23__device_stub__calcdataPdS_ii
.p2align 4, 0x90
.type _Z23__device_stub__calcdataPdS_ii,@function
_Z23__device_stub__calcdataPdS_ii: # @_Z23__device_stub__calcdataPdS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_o... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z8calcdataPdS_ii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R6, SR_CTAID.X ; /* 0x0000000000067919 */
/* 0x000e22000... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z8calcdataPdS_ii
.globl _Z8calcdataPdS_ii
.p2align 8
.type _Z8calcdataPdS_ii,@function
_Z8calcdataPdS_ii:
s_clause 0x1
s_load_b32 s2, s[0:1], 0x24
s_load_b64 s[4:5], s[0:1], 0x10
s_waitcnt lgkmcnt(0)
s_and_b32 s2, s2, 0xffff
s_de... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0001cad3_00000000-6_calcdata.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp... | .text
.file "calcdata.hip"
.globl _Z23__device_stub__calcdataPdS_ii # -- Begin function _Z23__device_stub__calcdataPdS_ii
.p2align 4, 0x90
.type _Z23__device_stub__calcdataPdS_ii,@function
_Z23__device_stub__calcdataPdS_ii: # @_Z23__device_stub__calcdataPdS_ii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_o... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | /*
Author: Jason He
Version: 1.0 20210605 Serial version.
Version: 2.0 20210607 CUDA version.
*/
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <math.h>
#include <time.h>
/*
This program solve the Stokes Flow problem- 36.1.2 in the textbook.
Inputs:
N, number of grid points in each direction.
mu, ... | .file "tmpxft_0000440b_00000000-6_stokes.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | /*
Author: Jason He
Version: 1.0 20210605 Serial version.
Version: 2.0 20210607 CUDA version.
*/
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <math.h>
#include <time.h>
/*
This program solve the Stokes Flow problem- 36.1.2 in the textbook.
Inputs:
N, number of grid points in each direction.
mu, ... | /*
Author: Jason He
Version: 1.0 20210605 Serial version.
Version: 2.0 20210607 CUDA version.
*/
#include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <math.h>
#include <time.h>
/*
This program solve the Stokes Flow problem- 36.1.2 in the textbook.
Inputs:
N, number of grid p... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | /*
Author: Jason He
Version: 1.0 20210605 Serial version.
Version: 2.0 20210607 CUDA version.
*/
#include <hip/hip_runtime.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <math.h>
#include <time.h>
/*
This program solve the Stokes Flow problem- 36.1.2 in the textbook.
Inputs:
N, number of grid p... | .text
.file "stokes.hip"
.globl _Z19__device_stub__initPdi # -- Begin function _Z19__device_stub__initPdi
.p2align 4, 0x90
.type _Z19__device_stub__initPdi,@function
_Z19__device_stub__initPdi: # @_Z19__device_stub__initPdi
.cfi_startproc
# %bb.0:
subq $88, %rsp
.cfi_def_cfa_offset 96
movq %rdi, 56(%rs... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "VectorAddition.cuh"
#include <cuda_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::size_t number_of_elements)
{
std::size_t i {blockDim.x * blockIdx.x + threadIdx.x};
if (i < number_of_elements)
{
C[i] = A[i] + B[i] + 0.0f;... | code for sm_80
Function : _ZN11CUDASamples12Introduction10vector_addEPKfS2_Pfm
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "VectorAddition.cuh"
#include <cuda_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::size_t number_of_elements)
{
std::size_t i {blockDim.x * blockIdx.x + threadIdx.x};
if (i < number_of_elements)
{
C[i] = A[i] + B[i] + 0.0f;... | .file "tmpxft_0003cef1_00000000-6_VectorAddition.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "VectorAddition.cuh"
#include <cuda_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::size_t number_of_elements)
{
std::size_t i {blockDim.x * blockIdx.x + threadIdx.x};
if (i < number_of_elements)
{
C[i] = A[i] + B[i] + 0.0f;... | #ifndef CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
#define CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
// For the CUDA runtime routines (prefixed with "cuda_").
#include <hip/hip_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::s... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #ifndef CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
#define CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
// For the CUDA runtime routines (prefixed with "cuda_").
#include <hip/hip_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::s... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #ifndef CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
#define CUDA_SAMPLES_INTRODUCTION_VECTOR_ADDITION_CUH
// For the CUDA runtime routines (prefixed with "cuda_").
#include <hip/hip_runtime.h>
namespace CUDASamples
{
namespace Introduction
{
__global__ void vector_add(
const float* A,
const float* B,
float* C,
std::s... | .text
.file "VectorAddition.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-p... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _ZN11CUDASamples12Introduction10vector_addEPKfS2_Pfm
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0003cef1_00000000-6_VectorAddition.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8... | .text
.file "VectorAddition.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-p... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} | code for sm_80
Function : _Z6gpuSumPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e280000002500 */
/*0020*... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} | .file "tmpxft_0016418f_00000000-6_gpuSum.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6gpuSumPiS_S_i
.globl _Z6gpuSumPiS_S_i
.p2align 8
.type _Z6gpuSumPiS_S_i,@function
_Z6gpuSumPiS_S_i:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b32 s8, s[0:1], 0x18
s_add_u32 s2, s0, 32
s_addc_u32 s3, s1, 0
s_waitcnt lgkm... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void gpuSum(int *a, int *b, int *c, int n) {
int idx = threadIdx.x + (blockIdx.x * blockDim.x);
while (idx < n) {
c[idx] = a[idx] + b[idx];
idx += blockDim.x * gridDim.x;
}
} | .text
.file "gpuSum.hip"
.globl _Z21__device_stub__gpuSumPiS_S_i # -- Begin function _Z21__device_stub__gpuSumPiS_S_i
.p2align 4, 0x90
.type _Z21__device_stub__gpuSumPiS_S_i,@function
_Z21__device_stub__gpuSumPiS_S_i: # @_Z21__device_stub__gpuSumPiS_S_i
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset ... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z6gpuSumPiS_S_i
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R0, SR_CTAID.X ; /* 0x0000000000007919 */
/* 0x000e280000002500 */
/*0020*... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z6gpuSumPiS_S_i
.globl _Z6gpuSumPiS_S_i
.p2align 8
.type _Z6gpuSumPiS_S_i,@function
_Z6gpuSumPiS_S_i:
s_clause 0x1
s_load_b32 s4, s[0:1], 0x2c
s_load_b32 s8, s[0:1], 0x18
s_add_u32 s2, s0, 32
s_addc_u32 s3, s1, 0
s_waitcnt lgkm... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0016418f_00000000-6_gpuSum.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rsp
.... | .text
.file "gpuSum.hip"
.globl _Z21__device_stub__gpuSumPiS_S_i # -- Begin function _Z21__device_stub__gpuSumPiS_S_i
.p2align 4, 0x90
.type _Z21__device_stub__gpuSumPiS_S_i,@function
_Z21__device_stub__gpuSumPiS_S_i: # @_Z21__device_stub__gpuSumPiS_S_i
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_offset ... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdio.h>
#include "cuda.h"
#define max(x,y) ((x) > (y)? (x) : (y))
#define min(x,y) ((x) < (y)? (x) : (y))
#define ceil(a,b) ((a) % (b) == 0 ? (a) / (b) : ((a) / (b)) + 1)
void check_error (const char* message) {
cudaError_t error = cudaGetLastError ();
if (error != cudaSuccess) {
printf ("CUDA error : %s, %s... | .file "tmpxft_001215ac_00000000-6_reordered-a.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <stdio.h>
#include "hip/hip_runtime.h"
#define max(x,y) ((x) > (y)? (x) : (y))
#define min(x,y) ((x) < (y)? (x) : (y))
#define ceil(a,b) ((a) % (b) == 0 ? (a) / (b) : ((a) / (b)) + 1)
void check_error (const char* message) {
hipError_t error = hipGetLastError ();
if (error != hipSuccess) {
printf ("CUDA error ... | .text
.file "reordered-a.hip"
.globl _Z11check_errorPKc # -- Begin function _Z11check_errorPKc
.p2align 4, 0x90
.type _Z11check_errorPKc,@function
_Z11check_errorPKc: # @_Z11check_errorPKc
.cfi_startproc
# %bb.0:
pushq %rbx
.cfi_def_cfa_offset 16
.cfi_offset %rbx, -16
movq %rdi, %rbx
ca... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_001215ac_00000000-6_reordered-a.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2061:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %... | .text
.file "reordered-a.hip"
.globl _Z11check_errorPKc # -- Begin function _Z11check_errorPKc
.p2align 4, 0x90
.type _Z11check_errorPKc,@function
_Z11check_errorPKc: # @_Z11check_errorPKc
.cfi_startproc
# %bb.0:
pushq %rbx
.cfi_def_cfa_offset 16
.cfi_offset %rbx, -16
movq %rdi, %rbx
ca... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | __global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((double)i * scale_real) - 2.25;
... | code for sm_80
Function : _Z9mandelgpuiiPii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ I2F.F64 R4, c[0x0][0x160] ; /* 0x0000580000047b12 */
/* 0x00... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | __global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((double)i * scale_real) - 2.25;
... | .file "tmpxft_000c1b44_00000000-6_mandelgpu.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | __global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((double)i * scale_real) - 2.25;
... | #include <hip/hip_runtime.h>
__global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((do... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
__global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((do... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z9mandelgpuiiPii
.globl _Z9mandelgpuiiPii
.p2align 8
.type _Z9mandelgpuiiPii,@function
_Z9mandelgpuiiPii:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_load_b32 s4, s[0:1], 0x4
v_and_b32_e32 v2, 0x3ff, v0... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
__global__ void mandelgpu(int disp_width, int disp_height, int *array, int max_iter) {
double scale_real, scale_imag;
double x, y, u, v, u2, v2;
int i, j, iter;
scale_real = 3.5 / (double)disp_width;
scale_imag = 3.5 / (double)disp_height;
i = blockIdx.x * blockDim.x + threadIdx.x;
x = ((do... | .text
.file "mandelgpu.hip"
.globl _Z24__device_stub__mandelgpuiiPii # -- Begin function _Z24__device_stub__mandelgpuiiPii
.p2align 4, 0x90
.type _Z24__device_stub__mandelgpuiiPii,@function
_Z24__device_stub__mandelgpuiiPii: # @_Z24__device_stub__mandelgpuiiPii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z9mandelgpuiiPii
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ I2F.F64 R4, c[0x0][0x160] ; /* 0x0000580000047b12 */
/* 0x00... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z9mandelgpuiiPii
.globl _Z9mandelgpuiiPii
.p2align 8
.type _Z9mandelgpuiiPii,@function
_Z9mandelgpuiiPii:
s_clause 0x2
s_load_b32 s2, s[0:1], 0x24
s_load_b32 s3, s[0:1], 0x10
s_load_b32 s4, s[0:1], 0x4
v_and_b32_e32 v2, 0x3ff, v0... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000c1b44_00000000-6_mandelgpu.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "mandelgpu.hip"
.globl _Z24__device_stub__mandelgpuiiPii # -- Begin function _Z24__device_stub__mandelgpuiiPii
.p2align 4, 0x90
.type _Z24__device_stub__mandelgpuiiPii,@function
_Z24__device_stub__mandelgpuiiPii: # @_Z24__device_stub__mandelgpuiiPii
.cfi_startproc
# %bb.0:
subq $120, %rsp
.cfi_def_cfa_... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <cuda_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
cudaError err = call; \
if( cudaSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LINE_... | code for sm_80 |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <cuda_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
cudaError err = call; \
if( cudaSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LINE_... | .file "tmpxft_0001b6d8_00000000-6_GPUPrimitive_Def.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2059:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <cuda_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
cudaError err = call; \
if( cudaSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LINE_... | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <hip/hip_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
hipError_t err = call; \
if( hipSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LI... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <hip/hip_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
hipError_t err = call; \
if( hipSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LI... | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #ifndef GPUPRIMITIVE_DEF_CU
#define GPUPRIMITIVE_DEF_CU
#include "stdlib.h"
#include <stdio.h>
#include <hip/hip_runtime.h>
//unsigned int gpuMemSize = 0;
# define CUDA_SAFE_CALL( call) { \
hipError_t err = call; \
if( hipSuccess != err) { \
fprintf(stderr, "Cuda error in file '%s' in line %i : %s.\n", \
__FILE__, __LI... | .text
.file "GPUPrimitive_Def.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80 | .text
.p2alignl 7, 3214868480
.fill 96, 4, 3214868480
.type __hip_cuid_,@object
.section .bss,"aw",@nobits
.globl __hip_cuid_
__hip_cuid_:
.byte 0
.size __hip_cuid_, 1
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm-project roc-6.3.2 25012 e5bf7... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0001b6d8_00000000-6_GPUPrimitive_Def.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2059:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... | .text
.file "GPUPrimitive_Def.hip"
.type __hip_cuid_,@object # @__hip_cuid_
.bss
.globl __hip_cuid_
__hip_cuid_:
.byte 0 # 0x0
.size __hip_cuid_, 1
.section ".linker-options","e",@llvm_linker_options
.ident "AMD clang version 18.0.0git (https://github.com/RadeonOpenCompute/llvm... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <stdlib.h>
#include <stdio.h>
#define NUM_BLOCKS 20
__device__ int **dataptr;
// Per-block pointer
__global__ void dynamic_allocmem()
{
// Only the first thread in the block does the allocation
// since we want only one allocation per block.
if (blockIdx.x == 0 && threadIdx.x == 0)
dataptr = (int**)m... | code for sm_80
Function : _Z7freememv
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ IMAD.MOV.U32 R2, RZ, RZ, c[0x4][0x10] ; /* 0x01000400ff027624 */
/... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <stdlib.h>
#include <stdio.h>
#define NUM_BLOCKS 20
__device__ int **dataptr;
// Per-block pointer
__global__ void dynamic_allocmem()
{
// Only the first thread in the block does the allocation
// since we want only one allocation per block.
if (blockIdx.x == 0 && threadIdx.x == 0)
dataptr = (int**)m... | .file "tmpxft_0005c162_00000000-6_allocation_persistence.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <stdlib.h>
#include <stdio.h>
#define NUM_BLOCKS 20
__device__ int **dataptr;
// Per-block pointer
__global__ void dynamic_allocmem()
{
// Only the first thread in the block does the allocation
// since we want only one allocation per block.
if (blockIdx.x == 0 && threadIdx.x == 0)
dataptr = (int**)m... | #include <hip/hip_runtime.h>
#include <stdlib.h>
#include <stdio.h>
#define NUM_BLOCKS 20
__device__ int **dataptr;
// Per-block pointer
__global__ void dynamic_allocmem()
{
// Only the first thread in the block does the allocation
// since we want only one allocation per block.
if (blockIdx.x == 0 && threadId... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdlib.h>
#include <stdio.h>
#define NUM_BLOCKS 20
__device__ int **dataptr;
// Per-block pointer
__global__ void dynamic_allocmem()
{
// Only the first thread in the block does the allocation
// since we want only one allocation per block.
if (blockIdx.x == 0 && threadId... | .text
.file "allocation_persistence.hip"
.globl _Z31__device_stub__dynamic_allocmemv # -- Begin function _Z31__device_stub__dynamic_allocmemv
.p2align 4, 0x90
.type _Z31__device_stub__dynamic_allocmemv,@function
_Z31__device_stub__dynamic_allocmemv: # @_Z31__device_stub__dynamic_allocmemv
.cfi_startproc
# %bb.0:
subq... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0005c162_00000000-6_allocation_persistence.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT... | .text
.file "allocation_persistence.hip"
.globl _Z31__device_stub__dynamic_allocmemv # -- Begin function _Z31__device_stub__dynamic_allocmemv
.p2align 4, 0x90
.type _Z31__device_stub__dynamic_allocmemv,@function
_Z31__device_stub__dynamic_allocmemv: # @_Z31__device_stub__dynamic_allocmemv
.cfi_startproc
# %bb.0:
subq... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx] += g_idata[idx + blockDim.x... | code for sm_80
Function : _Z16reduceUnrolling4PiS_j
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e220000002500 *... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx] += g_idata[idx + blockDim.x... | .file "tmpxft_0000a2b5_00000000-6_reduceUnrolling4.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx] += g_idata[idx + blockDim.x... | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx... |
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z16reduceUnrolling4PiS_j
.globl _Z16reduceUnrolling4PiS_j
.p2align 8
.type _Z16reduceUnrolling4PiS_j,@function
_Z16reduceUnrolling4PiS_j:
s_clause 0x2
s_load_b32 s3, s[0:1], 0x24
s_load_b32 s7, s[0:1], 0x10
s_load_b64 s[4:5], s[0:1], 0x0
s_mov... |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include "includes.h"
__global__ void reduceUnrolling4(int *g_idata, int *g_odata, unsigned int n){
unsigned int tid = threadIdx.x;
unsigned int idx = (4 * blockIdx.x) * blockDim.x + threadIdx.x;
int *idata = g_idata + (4 * blockIdx.x) * blockDim.x;
if(idx + 3 * blockDim.x < n){
g_idata[idx... | .text
.file "reduceUnrolling4.hip"
.globl _Z31__device_stub__reduceUnrolling4PiS_j # -- Begin function _Z31__device_stub__reduceUnrolling4PiS_j
.p2align 4, 0x90
.type _Z31__device_stub__reduceUnrolling4PiS_j,@function
_Z31__device_stub__reduceUnrolling4PiS_j: # @_Z31__device_stub__reduceUnrolling4PiS_j
.cfi_startproc
#... |
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly. | code for sm_80
Function : _Z16reduceUnrolling4PiS_j
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */
/* 0x000fe40000000f00 */
/*0010*/ S2R R4, SR_CTAID.X ; /* 0x0000000000047919 */
/* 0x000e220000002500 *... | .text
.amdgcn_target "amdgcn-amd-amdhsa--gfx1100"
.protected _Z16reduceUnrolling4PiS_j
.globl _Z16reduceUnrolling4PiS_j
.p2align 8
.type _Z16reduceUnrolling4PiS_j,@function
_Z16reduceUnrolling4PiS_j:
s_clause 0x2
s_load_b32 s3, s[0:1], 0x24
s_load_b32 s7, s[0:1], 0x10
s_load_b64 s[4:5], s[0:1], 0x0
s_mov... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_0000a2b5_00000000-6_reduceUnrolling4.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2029:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq ... | .text
.file "reduceUnrolling4.hip"
.globl _Z31__device_stub__reduceUnrolling4PiS_j # -- Begin function _Z31__device_stub__reduceUnrolling4PiS_j
.p2align 4, 0x90
.type _Z31__device_stub__reduceUnrolling4PiS_j,@function
_Z31__device_stub__reduceUnrolling4PiS_j: # @_Z31__device_stub__reduceUnrolling4PiS_j
.cfi_startproc
#... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly. | #include <cuda_runtime.h>
#include <stdio.h>
__global__ void kernel() {
int tid = threadIdx.x;
if (tid < 8) {
printf("inside the kernel\n");
}
else {
printf("outside the kernel\n");
}
}
int cuda(int a, int b) {
kernel<<<1, 10>>>();
cudaDeviceSynchronize();
return 0;
} | code for sm_80
Function : _Z6kernelv
.headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)"
/*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */
/* 0x000fe400078e00ff */
/*0010*/ S2R R0, SR_TID.X ; /* 0x0000000000007919 */
/* 0x000e240000002100 *... |
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly. | #include <cuda_runtime.h>
#include <stdio.h>
__global__ void kernel() {
int tid = threadIdx.x;
if (tid < 8) {
printf("inside the kernel\n");
}
else {
printf("outside the kernel\n");
}
}
int cuda(int a, int b) {
kernel<<<1, 10>>>();
cudaDeviceSynchronize();
return 0;
} | .file "tmpxft_000b80b7_00000000-6_cuda_test.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... |
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code. | #include <cuda_runtime.h>
#include <stdio.h>
__global__ void kernel() {
int tid = threadIdx.x;
if (tid < 8) {
printf("inside the kernel\n");
}
else {
printf("outside the kernel\n");
}
}
int cuda(int a, int b) {
kernel<<<1, 10>>>();
cudaDeviceSynchronize();
return 0;
} | #include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void kernel() {
int tid = threadIdx.x;
if (tid < 8) {
printf("inside the kernel\n");
}
else {
printf("outside the kernel\n");
}
}
int cuda(int a, int b) {
kernel<<<1, 10>>>();
hipDeviceSynchronize();
return 0;
} |
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly. | #include <hip/hip_runtime.h>
#include <stdio.h>
__global__ void kernel() {
int tid = threadIdx.x;
if (tid < 8) {
printf("inside the kernel\n");
}
else {
printf("outside the kernel\n");
}
}
int cuda(int a, int b) {
kernel<<<1, 10>>>();
hipDeviceSynchronize();
return 0;
} | .text
.file "cuda_test.hip"
.globl _Z21__device_stub__kernelv # -- Begin function _Z21__device_stub__kernelv
.p2align 4, 0x90
.type _Z21__device_stub__kernelv,@function
_Z21__device_stub__kernelv: # @_Z21__device_stub__kernelv
.cfi_startproc
# %bb.0:
subq $56, %rsp
.cfi_def_cfa_offset 64
leaq 32(%rsp),... |
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly. | .file "tmpxft_000b80b7_00000000-6_cuda_test.cudafe1.cpp"
.text
#APP
#NO_APP
.type _ZL26__cudaUnregisterBinaryUtilv, @function
_ZL26__cudaUnregisterBinaryUtilv:
.LFB2060:
.cfi_startproc
endbr64
subq $8, %rsp
.cfi_def_cfa_offset 16
movq _ZL20__cudaFatCubinHandle(%rip), %rdi
call __cudaUnregisterFatBinary@PLT
addq $8, %rs... | .text
.file "cuda_test.hip"
.globl _Z21__device_stub__kernelv # -- Begin function _Z21__device_stub__kernelv
.p2align 4, 0x90
.type _Z21__device_stub__kernelv,@function
_Z21__device_stub__kernelv: # @_Z21__device_stub__kernelv
.cfi_startproc
# %bb.0:
subq $56, %rsp
.cfi_def_cfa_offset 64
leaq 32(%rsp),... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.