Dataset Viewer
Auto-converted to Parquet Duplicate
system
stringclasses
7 values
input
stringlengths
10
94.5k
output
stringlengths
14
55.5k
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly.
#include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
code for sm_80 Function : _Z33PadMatrixInLeadingDimensionKernelPKaPaii .headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)" /*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */ /* 0x000fe400078e00ff */ /*0010*/ S2R R0, SR_TID.X ; /* 0x000000000...
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly.
#include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
.file "tmpxft_00097a77_00000000-6_PadMatrixInLeadingDimensionKernel.cudafe1.cpp" .text #APP #NO_APP .type _ZL26__cudaUnregisterBinaryUtilv, @function _ZL26__cudaUnregisterBinaryUtilv: .LFB2029: .cfi_startproc endbr64 subq $8, %rsp .cfi_def_cfa_offset 16 movq _ZL20__cudaFatCubinHandle(%rip), %rdi call __cudaUnregisterFa...
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code.
#include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
#include <hip/hip_runtime.h> #include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly.
#include <hip/hip_runtime.h> #include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
.text .amdgcn_target "amdgcn-amd-amdhsa--gfx1100" .protected _Z33PadMatrixInLeadingDimensionKernelPKaPaii .globl _Z33PadMatrixInLeadingDimensionKernelPKaPaii .p2align 8 .type _Z33PadMatrixInLeadingDimensionKernelPKaPaii,@function _Z33PadMatrixInLeadingDimensionKernelPKaPaii: s_load_b32 s2, s[0:1], 0x10 s_mov_b32 s3, ...
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly.
#include <hip/hip_runtime.h> #include "includes.h" __global__ void PadMatrixInLeadingDimensionKernel(const int8_t* src, int8_t* dst, int col_src, int col_dst) { for (int32_t i = threadIdx.x; i < col_src; i += blockDim.x) { *(dst + blockIdx.x * col_dst + i) = *(src + blockIdx.x * col_src + i); } }
.text .file "PadMatrixInLeadingDimensionKernel.hip" .globl _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii # -- Begin function _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii .p2align 4, 0x90 .type _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii,@function _Z48__device_stub__PadMatrix...
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly.
code for sm_80 Function : _Z33PadMatrixInLeadingDimensionKernelPKaPaii .headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)" /*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */ /* 0x000fe400078e00ff */ /*0010*/ S2R R0, SR_TID.X ; /* 0x000000000...
.text .amdgcn_target "amdgcn-amd-amdhsa--gfx1100" .protected _Z33PadMatrixInLeadingDimensionKernelPKaPaii .globl _Z33PadMatrixInLeadingDimensionKernelPKaPaii .p2align 8 .type _Z33PadMatrixInLeadingDimensionKernelPKaPaii,@function _Z33PadMatrixInLeadingDimensionKernelPKaPaii: s_load_b32 s2, s[0:1], 0x10 s_mov_b32 s3, ...
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly.
.file "tmpxft_00097a77_00000000-6_PadMatrixInLeadingDimensionKernel.cudafe1.cpp" .text #APP #NO_APP .type _ZL26__cudaUnregisterBinaryUtilv, @function _ZL26__cudaUnregisterBinaryUtilv: .LFB2029: .cfi_startproc endbr64 subq $8, %rsp .cfi_def_cfa_offset 16 movq _ZL20__cudaFatCubinHandle(%rip), %rdi call __cudaUnregisterFa...
.text .file "PadMatrixInLeadingDimensionKernel.hip" .globl _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii # -- Begin function _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii .p2align 4, 0x90 .type _Z48__device_stub__PadMatrixInLeadingDimensionKernelPKaPaii,@function _Z48__device_stub__PadMatrix...
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly.
#include "cuda_runtime.h" #include "device_launch_parameters.h" #include <cuda.h> #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, i...
code for sm_80 Function : _Z7MatMultPfS_S_ii .headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)" /*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */ /* 0x000fe40000000f00 */ /*0010*/ IABS R5, c[0x0][0x17c] ; /* 0x00005f0000057a13 */ /* 0x000fe20000000000 */ /...
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly.
#include "cuda_runtime.h" #include "device_launch_parameters.h" #include <cuda.h> #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, i...
.file "tmpxft_00139677_00000000-6_kernel.cudafe1.cpp" .text #APP #NO_APP .type _ZL26__cudaUnregisterBinaryUtilv, @function _ZL26__cudaUnregisterBinaryUtilv: .LFB2062: .cfi_startproc endbr64 subq $8, %rsp .cfi_def_cfa_offset 16 movq _ZL20__cudaFatCubinHandle(%rip), %rdi call __cudaUnregisterFatBinary@PLT addq $8, %rsp ....
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code.
#include "cuda_runtime.h" #include "device_launch_parameters.h" #include <cuda.h> #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, i...
#include "hip/hip_runtime.h" #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, int tileWidth) { __shared__ float Mds[TILE_WIDTH][TILE...
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly.
#include "hip/hip_runtime.h" #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, int tileWidth) { __shared__ float Mds[TILE_WIDTH][TILE...
.text .amdgcn_target "amdgcn-amd-amdhsa--gfx1100" .protected _Z7MatMultPfS_S_ii .globl _Z7MatMultPfS_S_ii .p2align 8 .type _Z7MatMultPfS_S_ii,@function _Z7MatMultPfS_S_ii: s_load_b64 s[8:9], s[0:1], 0x18 v_and_b32_e32 v3, 0x3ff, v0 v_bfe_u32 v2, v0, 10, 10 s_mov_b32 s10, 0 s_waitcnt lgkmcnt(0...
You are a helpful coding assistant specialized in converting HIP source code to AMD host assembly.
#include "hip/hip_runtime.h" #include <stdio.h> #include <stdlib.h> #include <math.h> // thread block size #define BLOCKDIM 16 #define TILE_WIDTH 2 // threshold #define TOLERANCE 0.01 float absf(float n); __global__ void MatMult(float *a, float *b, float *c, int N, int tileWidth) { __shared__ float Mds[TILE_WIDTH][TILE...
.text .file "kernel.hip" .globl _Z22__device_stub__MatMultPfS_S_ii # -- Begin function _Z22__device_stub__MatMultPfS_S_ii .p2align 4, 0x90 .type _Z22__device_stub__MatMultPfS_S_ii,@function _Z22__device_stub__MatMultPfS_S_ii: # @_Z22__device_stub__MatMultPfS_S_ii .cfi_startproc # %bb.0: subq $120, %rsp .cfi_def_cfa...
You are a helpful coding assistant specialized in converting from CUDA device assembly to AMD device assembly.
code for sm_80 Function : _Z7MatMultPfS_S_ii .headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)" /*0000*/ MOV R1, c[0x0][0x28] ; /* 0x00000a0000017a02 */ /* 0x000fe40000000f00 */ /*0010*/ IABS R5, c[0x0][0x17c] ; /* 0x00005f0000057a13 */ /* 0x000fe20000000000 */ /...
.text .amdgcn_target "amdgcn-amd-amdhsa--gfx1100" .protected _Z7MatMultPfS_S_ii .globl _Z7MatMultPfS_S_ii .p2align 8 .type _Z7MatMultPfS_S_ii,@function _Z7MatMultPfS_S_ii: s_load_b64 s[8:9], s[0:1], 0x18 v_and_b32_e32 v3, 0x3ff, v0 v_bfe_u32 v2, v0, 10, 10 s_mov_b32 s10, 0 s_waitcnt lgkmcnt(0...
You are a helpful coding assistant specialized in converting from CUDA host assembly to AMD host assembly.
.file "tmpxft_00139677_00000000-6_kernel.cudafe1.cpp" .text #APP #NO_APP .type _ZL26__cudaUnregisterBinaryUtilv, @function _ZL26__cudaUnregisterBinaryUtilv: .LFB2062: .cfi_startproc endbr64 subq $8, %rsp .cfi_def_cfa_offset 16 movq _ZL20__cudaFatCubinHandle(%rip), %rdi call __cudaUnregisterFatBinary@PLT addq $8, %rsp ....
.text .file "kernel.hip" .globl _Z22__device_stub__MatMultPfS_S_ii # -- Begin function _Z22__device_stub__MatMultPfS_S_ii .p2align 4, 0x90 .type _Z22__device_stub__MatMultPfS_S_ii,@function _Z22__device_stub__MatMultPfS_S_ii: # @_Z22__device_stub__MatMultPfS_S_ii .cfi_startproc # %bb.0: subq $120, %rsp .cfi_def_cfa...
You are a helpful coding assistant specialized in converting CUDA source code to CUDA device assembly.
#include "includes.h" __device__ int sign(DECNUM x) { return((x > 0.0f) - (x < 0.0f)); } __device__ int mminus2(int ix, int nx) { int xminus; if (ix <= 1) { xminus = 0; } else { xminus = ix - 2; } return(xminus); } __device__ int pplus(int ix, int nx) { int xplus; if (ix == nx - 1) { xplus = nx - 1; } else { xplus = ix...
code for sm_80 Function : _Z8vvlatbndiiPf .headerflags @"EF_CUDA_TEXMODE_UNIFIED EF_CUDA_64BIT_ADDRESS EF_CUDA_SM80 EF_CUDA_VIRTUAL_SM(EF_CUDA_SM80)" /*0000*/ IMAD.MOV.U32 R1, RZ, RZ, c[0x0][0x28] ; /* 0x00000a00ff017624 */ /* 0x000fe400078e00ff */ /*0010*/ S2R R7, SR_CTAID.Y ; /* 0x0000000000077919 */ /* 0x000e2800000...
You are a helpful coding assistant specialized in converting CUDA source code to CUDA host assembly.
#include "includes.h" __device__ int sign(DECNUM x) { return((x > 0.0f) - (x < 0.0f)); } __device__ int mminus2(int ix, int nx) { int xminus; if (ix <= 1) { xminus = 0; } else { xminus = ix - 2; } return(xminus); } __device__ int pplus(int ix, int nx) { int xplus; if (ix == nx - 1) { xplus = nx - 1; } else { xplus = ix...
.file "tmpxft_0014341b_00000000-6_vvlatbnd.cudafe1.cpp" .text #APP #NO_APP .type _ZL26__cudaUnregisterBinaryUtilv, @function _ZL26__cudaUnregisterBinaryUtilv: .LFB2033: .cfi_startproc endbr64 subq $8, %rsp .cfi_def_cfa_offset 16 movq _ZL20__cudaFatCubinHandle(%rip), %rdi call __cudaUnregisterFatBinary@PLT addq $8, %rsp...
You are a helpful coding assistant specialized in converting CUDA source code to HIP/AMD source code.
#include "includes.h" __device__ int sign(DECNUM x) { return((x > 0.0f) - (x < 0.0f)); } __device__ int mminus2(int ix, int nx) { int xminus; if (ix <= 1) { xminus = 0; } else { xminus = ix - 2; } return(xminus); } __device__ int pplus(int ix, int nx) { int xplus; if (ix == nx - 1) { xplus = nx - 1; } else { xplus = ix...
#include <hip/hip_runtime.h> #include "includes.h" __device__ int sign(DECNUM x) { return((x > 0.0f) - (x < 0.0f)); } __device__ int mminus2(int ix, int nx) { int xminus; if (ix <= 1) { xminus = 0; } else { xminus = ix - 2; } return(xminus); } __device__ int pplus(int ix, int nx) { int xplus; if (ix == nx - 1) { xplus ...
You are a helpful coding assistant specialized in converting HIP source code to AMD device assembly.
#include <hip/hip_runtime.h> #include "includes.h" __device__ int sign(DECNUM x) { return((x > 0.0f) - (x < 0.0f)); } __device__ int mminus2(int ix, int nx) { int xminus; if (ix <= 1) { xminus = 0; } else { xminus = ix - 2; } return(xminus); } __device__ int pplus(int ix, int nx) { int xplus; if (ix == nx - 1) { xplus ...
.text .amdgcn_target "amdgcn-amd-amdhsa--gfx1100" .protected _Z8vvlatbndiiPf .globl _Z8vvlatbndiiPf .p2align 8 .type _Z8vvlatbndiiPf,@function _Z8vvlatbndiiPf: s_clause 0x1 s_load_b32 s2, s[0:1], 0x1c s_load_b64 s[4:5], s[0:1], 0x0 v_and_b32_e32 v1, 0x3ff, v0 v_bfe_u32 v2, v0, 10, 10 s_...
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
9