| |
| |
| |
| |
| |
| |
| |
| |
|
|
| #include <unistd.h> |
| #include <stdio.h> |
| #include <time.h> |
| #include <sys/time.h> |
| #include <stdlib.h> |
| #include <stdarg.h> |
| #include <string.h> |
| #include <cuda.h> |
| #include "../../../common/cupti_add.h" |
| #include "../../../common/cpu_timestamps.h" |
|
|
| #include <cooperative_groups.h> |
| #include <cooperative_groups/memcpy_async.h> |
|
|
| using namespace nvcuda::experimental; |
|
|
| #define PREFETCH_COUNT 2 |
|
|
| #define SMALL_FLOAT_VAL 0.00000001f |
|
|
| double rtclock() |
| { |
| struct timezone Tzp; |
| struct timeval Tp; |
| uint64_t stat; |
| stat = gettimeofday(&Tp, &Tzp); |
| if (stat != 0) |
| printf("Error return from gettimeofday: %d", stat); |
| return (Tp.tv_sec + Tp.tv_usec * 1.0e-6); |
| } |
|
|
| float absVal(float a) |
| { |
| if (a < 0) |
| { |
| return (a * -1); |
| } |
| else |
| { |
| return a; |
| } |
| } |
|
|
| float percentDiff(double val1, double val2) |
| { |
| if ((absVal(val1) < 0.01) && (absVal(val2) < 0.01)) |
| { |
| return 0.0f; |
| } |
|
|
| else |
| { |
| return 100.0f * (absVal(absVal(val1 - val2) / absVal(val1 + SMALL_FLOAT_VAL))); |
| } |
| } |
|
|
| |
| #define PERCENT_DIFF_ERROR_THRESHOLD 0.05 |
|
|
| |
| #define SIZE 40960 |
| uint64_t NI; |
| uint64_t NJ; |
|
|
| |
| #define DIM_THREAD_BLOCK 256 |
|
|
| #define BATCH_SIZE 16 |
|
|
| |
| #define ALPHA 1.1f |
| #define BETA 1.1f |
|
|
| |
| typedef float DATA_TYPE; |
| |
|
|
| void gemv(DATA_TYPE *A, DATA_TYPE *B, DATA_TYPE *C) |
| { |
| uint64_t i, j; |
|
|
| for (i = 0; i < NI; i++) |
| { |
| C[i] *= BETA; |
| for (j = 0; j < NJ; j++) |
| { |
| C[i] += ALPHA * A[i * NJ + j] * B[j]; |
| } |
| } |
| } |
|
|
| void init(DATA_TYPE *A, DATA_TYPE *B, DATA_TYPE *C, DATA_TYPE *C_ref) |
| { |
| uint64_t i, j; |
|
|
| for (i = 0; i < NI; i++) |
| for (j = 0; j < NJ; j++) |
| A[i * NJ + j] = ((DATA_TYPE)i * j) / NI; |
|
|
| for (j = 0; j < NJ; j++) |
| B[j] = ((DATA_TYPE)j + 1) / NJ; |
|
|
| for (i = 0; i < NI; i++) |
| { |
| C[i] = ((DATA_TYPE)i + 2) / NI; |
| C_ref[i] = ((DATA_TYPE)i + 2) / NI; |
| } |
| } |
|
|
| void compareResults(DATA_TYPE *C, DATA_TYPE *C_outputFromGpu) |
| { |
| uint64_t i, fail; |
| fail = 0; |
|
|
| |
| for (i = 0; i < NI; i++) |
| { |
| if (percentDiff(C[i], C_outputFromGpu[i]) > PERCENT_DIFF_ERROR_THRESHOLD) |
| { |
| fail++; |
| printf("%d, GPU is %f, CPU is %f.\n", i, C[i], C_outputFromGpu[i]); |
| } |
| } |
|
|
| |
| printf("Non-Matching CPU-GPU Outputs Beyond Error Threshold of %4.2f Percent: %d\n", PERCENT_DIFF_ERROR_THRESHOLD, fail); |
| } |
|
|
| __global__ void gemv_kernel(DATA_TYPE *a, DATA_TYPE *b, DATA_TYPE *c, uint64_t NI, uint64_t NJ) |
| { |
| uint64_t row = blockIdx.x * blockDim.x + threadIdx.x; |
| uint64_t tx = threadIdx.x; |
|
|
| __shared__ DATA_TYPE s_b[DIM_THREAD_BLOCK][BATCH_SIZE]; |
|
|
| DATA_TYPE tmp = BETA * c[row]; |
| __syncthreads(); |
|
|
| uint64_t tile = 0; |
| uint64_t end_tile = NJ / BATCH_SIZE; |
|
|
| for (; tile < end_tile; tile += 1) |
| { |
| uint64_t base_index = tile * BATCH_SIZE; |
| for (uint64_t k = 0; k < BATCH_SIZE; k++) |
| { |
| s_b[tx][k] = b[base_index + k]; |
| } |
| __syncthreads(); |
|
|
| for (uint64_t k = 0; k < BATCH_SIZE; k++) |
| { |
| tmp += ALPHA * a[row * NJ + base_index + k] * s_b[tx][k]; |
| } |
| __syncthreads(); |
| } |
| c[row] = tmp; |
| } |
|
|
| void gemvCuda(DATA_TYPE *A, DATA_TYPE *B, DATA_TYPE *C, DATA_TYPE *A_gpu, DATA_TYPE *B_gpu, DATA_TYPE *C_gpu) |
| { |
| double t_start, t_end; |
|
|
| dim3 block(DIM_THREAD_BLOCK); |
| dim3 grid(NI / (DIM_THREAD_BLOCK)); |
|
|
| |
| cudaMemcpy(A_gpu, A, sizeof(DATA_TYPE) * NI * NJ, cudaMemcpyHostToDevice); |
| cudaMemcpy(B_gpu, B, sizeof(DATA_TYPE) * NJ, cudaMemcpyHostToDevice); |
| cudaMemcpy(C_gpu, C, sizeof(DATA_TYPE) * NI, cudaMemcpyHostToDevice); |
| gemv_kernel<<<grid, block>>>(A_gpu, B_gpu, C_gpu, NI, NJ); |
| cudaDeviceSynchronize(); |
| cudaMemcpy(C, C_gpu, sizeof(DATA_TYPE) * NI, cudaMemcpyDeviceToHost); |
| |
|
|
| |
| } |
|
|
| int main(int argc, char *argv[]) |
| { |
| uint64_t start_tsc = rdtsc(); |
| uint64_t start_tsp = rdtsp(); |
| printf("start_tsc %lu start_tsp %lu\n", start_tsc, start_tsp); |
| if (argc >= 3) |
| { |
| NI = atoll(argv[1]); |
| NJ = atoll(argv[2]); |
| } |
| else |
| { |
| NI = SIZE; |
| NJ = SIZE; |
| } |
|
|
| double t_start, t_end; |
|
|
| DATA_TYPE *A; |
| DATA_TYPE *B; |
| DATA_TYPE *C; |
| DATA_TYPE *C_ref; |
|
|
| DATA_TYPE *A_gpu; |
| DATA_TYPE *B_gpu; |
| DATA_TYPE *C_gpu; |
|
|
| A = (DATA_TYPE *)malloc(NI * NJ * sizeof(DATA_TYPE)); |
| B = (DATA_TYPE *)malloc(NJ * sizeof(DATA_TYPE)); |
| C = (DATA_TYPE *)malloc(NI * sizeof(DATA_TYPE)); |
| C_ref = (DATA_TYPE *)malloc(NI * sizeof(DATA_TYPE)); |
|
|
| |
| |
| |
|
|
| init(A, B, C, C_ref); |
|
|
| GPU_argv_init(); |
| initTrace(); |
| startCPU(); |
|
|
| cudaMalloc(&A_gpu, sizeof(DATA_TYPE) * NI * NJ); |
| cudaMalloc(&B_gpu, sizeof(DATA_TYPE) * NJ); |
| cudaMalloc(&C_gpu, sizeof(DATA_TYPE) * NI); |
|
|
| gemvCuda(A, B, C, A_gpu, B_gpu, C_gpu); |
|
|
| cudaFree(A_gpu); |
| cudaFree(B_gpu); |
| cudaFree(C_gpu); |
| endCPU(); |
| finiTrace(); |
|
|
| |
| |
| |
| |
|
|
| |
| free(A); |
| free(B); |
| free(C); |
| free(C_ref); |
| return 0; |
| } |
|
|