23#ifdef MLIR_ENABLE_CUDA_CUSPARSE
25#ifdef MLIR_ENABLE_CUDA_CUSPARSELT
26#include "cusparseLt.h"
32#define MLIR_CUDA_WRAPPERS_EXPORT __declspec(dllexport)
34#define MLIR_CUDA_WRAPPERS_EXPORT __attribute__((visibility("default")))
37#define CUDA_REPORT_IF_ERROR(expr) \
38 [](CUresult result) { \
41 const char *name = nullptr; \
42 cuGetErrorName(result, &name); \
45 fprintf(stderr, "'%s' failed with '%s'\n", #expr, name); \
53 return result == CUDA_ERROR_DEINITIALIZED ||
54 result == CUDA_ERROR_CONTEXT_IS_DESTROYED;
60#define CUDA_REPORT_IF_ERROR_IGNORE_SHUTDOWN(expr) \
61 [](CUresult result) { \
62 if (!result || isCudaContextShutdownError(result)) \
64 const char *name = nullptr; \
65 cuGetErrorName(result, &name); \
68 fprintf(stderr, "'%s' failed with '%s'\n", #expr, name); \
71#define CUSPARSE_REPORT_IF_ERROR(expr) \
73 cusparseStatus_t status = (expr); \
74 if (status != CUSPARSE_STATUS_SUCCESS) { \
75 fprintf(stderr, "cuSPARSE '%s' failed with '%s'\n", #expr, \
76 cusparseGetErrorString(status)); \
84 const char *kDebugEnvironmentVariable =
"MLIR_CUDA_DEBUG";
85 static bool isEnabled = getenv(kDebugEnvironmentVariable) !=
nullptr;
89#define debug_print(fmt, ...) \
91 if (isDebugEnabled()) \
92 fprintf(stderr, "%s:%d:%s(): " fmt, "CudaRuntimeWrappers.cpp", __LINE__, \
93 __func__, __VA_ARGS__); \
111 static CUcontext context = [] {
126#ifdef MLIR_ENABLE_CUDA_CUSPARSE
131static cusparseHandle_t cusparse_env =
nullptr;
133#ifdef MLIR_ENABLE_CUDA_CUSPARSELT
136static cusparseLtHandle_t cusparseLt_env;
137static bool cusparseLt_initiated =
false;
145 CUmodule module =
nullptr;
153 CUmodule module =
nullptr;
154 char jitErrorBuffer[4096] = {0};
155 CUjit_option jitOptions[] = {CU_JIT_ERROR_LOG_BUFFER,
156 CU_JIT_ERROR_LOG_BUFFER_SIZE_BYTES,
157 CU_JIT_OPTIMIZATION_LEVEL};
158 void *jitOptionsVals[] = {jitErrorBuffer,
159 reinterpret_cast<void *
>(
sizeof(jitErrorBuffer)),
160 reinterpret_cast<void *
>(optLevel)};
163 cuModuleLoadDataEx(&module, data, 3, jitOptions, jitOptionsVals);
165 fprintf(stderr,
"JIT compilation failed with: '%s'\n", jitErrorBuffer);
177 CUfunction function =
nullptr;
188 intptr_t blockZ, int32_t smem, CUstream stream,
void **params,
189 void **extra,
size_t ) {
193 int32_t maxShmem = 0;
197 &maxShmem, CU_DEVICE_ATTRIBUTE_MAX_SHARED_MEMORY_PER_BLOCK_OPTIN,
199 if (maxShmem < smem) {
201 "Requested shared memory (%dkb) is larger than maximum allowed "
202 "shared memory (%dkb) for this device\n",
206 function, CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES, smem));
209 "threads: %ld, %ld, %ld, "
211 gridX, gridY, gridZ, blockX, blockY, blockZ, smem);
213 blockY, blockZ, smem, stream, params,
219 CUstream stream =
nullptr;
236 cuStreamWaitEvent(stream, event, 0));
241 CUevent
event =
nullptr;
264 return reinterpret_cast<void *
>(ptr);
268 cuMemAllocManaged(&ptr, sizeBytes, CU_MEM_ATTACH_GLOBAL));
269 return reinterpret_cast<void *
>(ptr);
272 return reinterpret_cast<void *
>(ptr);
281mgpuMemcpy(
void *dst,
void *src,
size_t sizeBytes, CUstream stream) {
283 reinterpret_cast<CUdeviceptr
>(src),
288mgpuMemset32(
void *dst,
unsigned int value,
size_t count, CUstream stream) {
290 value, count, stream));
294mgpuMemset16(
void *dst,
unsigned short value,
size_t count, CUstream stream) {
296 value, count, stream));
324 for (
int64_t i = rank - 1, runningStride = 1; i >= 0; i--) {
325 denseStrides[i] = runningStride;
326 runningStride *= sizes[i];
328 uint64_t sizeBytes = sizes[0] * denseStrides[0] * elementSizeBytes;
329 int64_t *strides = &sizes[rank];
331 for (
unsigned i = 0; i < rank; ++i)
332 assert(strides[i] == denseStrides[i] &&
333 "Mismatch in computed dense strides");
335 auto *ptr = descriptor->
data + descriptor->
offset * elementSizeBytes;
351 auto *ptr = descriptor->
data + descriptor->
offset * elementSizeBytes;
363#if (CUDA_VERSION >= 12000)
369 void **params,
void **extra) {
372 int32_t maxShmem = 0;
376 &maxShmem, CU_DEVICE_ATTRIBUTE_MAX_SHARED_MEMORY_PER_BLOCK_OPTIN,
378 if (maxShmem < smem) {
380 "Requested shared memory (%dkb) is larger than maximum allowed "
381 "shared memory (%dkb) for this device\n",
385 function, CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES, smem));
388 CUlaunchConfig config;
389 config.gridDimX = gridX;
390 config.gridDimY = gridY;
391 config.gridDimZ = gridZ;
392 config.blockDimX = blockX;
393 config.blockDimY = blockY;
394 config.blockDimZ = blockZ;
395 config.sharedMemBytes = smem;
396 config.hStream = stream;
398 CUlaunchAttribute launchAttrs[3];
401 bool hasCluster = clusterX > 0 && clusterY > 0 && clusterZ > 0;
403 launchAttrs[numAttrs].id = CU_LAUNCH_ATTRIBUTE_CLUSTER_DIMENSION;
404 launchAttrs[numAttrs].value.clusterDim.x = clusterX;
405 launchAttrs[numAttrs].value.clusterDim.y = clusterY;
406 launchAttrs[numAttrs].value.clusterDim.z = clusterZ;
409 launchAttrs[numAttrs].id =
410 CU_LAUNCH_ATTRIBUTE_CLUSTER_SCHEDULING_POLICY_PREFERENCE;
411 launchAttrs[numAttrs].value.clusterSchedulingPolicyPreference =
412 CU_CLUSTER_SCHEDULING_POLICY_SPREAD;
416 launchAttrs[numAttrs].id = CU_LAUNCH_ATTRIBUTE_COOPERATIVE;
417 launchAttrs[numAttrs].value.cooperative = 1;
420 config.numAttrs = numAttrs;
421 config.attrs = launchAttrs;
423 debug_print(
"Launching cooperative kernel (cluster=%d), "
425 "threads: %ld, %ld, %ld, "
427 hasCluster, gridX, gridY, gridZ, blockX, blockY, blockZ, smem);
436 CUstream stream,
void **params,
void **extra,
size_t ) {
440 int32_t maxShmem = 0;
444 &maxShmem, CU_DEVICE_ATTRIBUTE_MAX_SHARED_MEMORY_PER_BLOCK_OPTIN,
446 if (maxShmem < smem) {
448 "Requested shared memory (%dkb) is larger than maximum allowed "
449 "shared memory (%dkb) for this device\n",
453 function, CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES, smem));
455 CUlaunchConfig config;
456 config.gridDimX = gridX;
457 config.gridDimY = gridY;
458 config.gridDimZ = gridZ;
459 config.blockDimX = blockX;
460 config.blockDimY = blockY;
461 config.blockDimZ = blockZ;
462 config.sharedMemBytes = smem;
463 config.hStream = stream;
464 CUlaunchAttribute launchAttr[2];
465 launchAttr[0].id = CU_LAUNCH_ATTRIBUTE_CLUSTER_DIMENSION;
466 launchAttr[0].value.clusterDim.x = clusterX;
467 launchAttr[0].value.clusterDim.y = clusterY;
468 launchAttr[0].value.clusterDim.z = clusterZ;
469 launchAttr[1].id = CU_LAUNCH_ATTRIBUTE_CLUSTER_SCHEDULING_POLICY_PREFERENCE;
470 launchAttr[1].value.clusterSchedulingPolicyPreference =
471 CU_CLUSTER_SCHEDULING_POLICY_SPREAD;
473 config.attrs = launchAttr;
476 "cluster: %ld, %ld, %ld, "
478 "threads: %ld, %ld, %ld, "
480 clusterX, clusterY, clusterZ, gridX, gridY, gridZ, blockX, blockY,
487 CUtensorMap *tensorMap,
488 CUtensorMapDataType tensorDataType,
489 cuuint32_t tensorRank,
491 const cuuint64_t *globalDim,
492 const cuuint64_t *globalStrides,
493 const cuuint32_t *boxDim,
494 const cuuint32_t *elementStrides,
495 CUtensorMapInterleave interleave,
496 CUtensorMapSwizzle swizzle,
497 CUtensorMapL2promotion l2Promotion,
498 CUtensorMapFloatOOBfill oobFill
502 tensorMap, tensorDataType, tensorRank, globalAddress, globalDim,
503 globalStrides, boxDim, elementStrides, interleave, swizzle, l2Promotion,
508 "globalDim[5]: %zu, %zu, %zu, %zu, %zu\n"
509 "globalStrides[5]: %zu, %zu, %zu, %zu, %zu\n"
510 "boxDim[5]: %u, %u, %u, %u, %u\n"
511 "elementStrides[5]: %u, %u, %u, %u, %u\n"
516 (
void *)&tensorMap, tensorDataType, tensorRank, globalDim[0],
517 globalDim[1], globalDim[2], globalDim[3], globalDim[4],
518 globalStrides[0], globalStrides[1], globalStrides[2],
519 globalStrides[3], globalStrides[4], boxDim[0], boxDim[1],
520 boxDim[2], boxDim[3], boxDim[4], elementStrides[0],
521 elementStrides[1], elementStrides[2], elementStrides[3],
522 elementStrides[4], interleave, swizzle, l2Promotion, oobFill);
526void mgpuGetMemRefDataAndShape(
void *rawDescriptor,
char **addr,
527 uint64_t *globalDim, uint64_t *globalStrides,
528 const CUtensorMapDataType tensorDataType) {
531 *addr = descriptor->data;
532 for (
int i = 0; i < Rank; ++i) {
533 globalDim[i] =
static_cast<uint64_t
>(descriptor->sizes[Rank - i - 1]);
535 static constexpr int elementSizeInBytes[] = {1, 2, 4, 4, 8, 8, 2,
537 for (
int i = 0; i < Rank - 1; ++i) {
538 globalStrides[i] =
static_cast<uint64_t
>(
539 descriptor->strides[Rank - i - 2] * elementSizeInBytes[tensorDataType]);
545 void *rankedDescriptor,
546 const CUtensorMapDataType tensorDataType,
547 CUtensorMapInterleave interleave,
548 CUtensorMapSwizzle swizzle,
549 CUtensorMapL2promotion l2Promotion,
550 CUtensorMapFloatOOBfill oobFill,
553 CUtensorMap tensorMap;
555 uint32_t boxDim[5] = {1, 1, 1, 1, 1}, elementStrides[5] = {1, 1, 1, 1, 1};
556 uint64_t globalDim[5] = {1, 1, 1, 1, 1}, globalStrides[5] = {0};
557 uint32_t tensorRank32 = uint32_t(tensorRank);
559 char *globalAddress =
nullptr;
560 switch (tensorRank) {
562 mgpuGetMemRefDataAndShape<1>(rankedDescriptor, &globalAddress, globalDim,
563 globalStrides, tensorDataType);
566 mgpuGetMemRefDataAndShape<2>(rankedDescriptor, &globalAddress, globalDim,
567 globalStrides, tensorDataType);
570 mgpuGetMemRefDataAndShape<3>(rankedDescriptor, &globalAddress, globalDim,
571 globalStrides, tensorDataType);
574 mgpuGetMemRefDataAndShape<4>(rankedDescriptor, &globalAddress, globalDim,
575 globalStrides, tensorDataType);
578 mgpuGetMemRefDataAndShape<5>(rankedDescriptor, &globalAddress, globalDim,
579 globalStrides, tensorDataType);
584 "'mgpuTensorMapEncodeTiledMemref' failed with 'rank is too high'\n");
588 for (
int64_t r = 0; r < tensorRank; ++r) {
589 boxDim[r] =
static_cast<uint32_t
>(inputBoxDims[tensorRank - r - 1]);
593 mgpuTensorMapEncodeTiled(&tensorMap, tensorDataType, tensorRank32,
594 globalAddress, globalDim, globalStrides, boxDim,
595 elementStrides, interleave, swizzle, l2Promotion,
598 CUdeviceptr dTensorMap;
601 reinterpret_cast<CUdeviceptr
>(&tensorMap),
602 sizeof(CUtensorMap)));
603 return reinterpret_cast<void *
>(dTensorMap);
607#ifdef MLIR_ENABLE_CUDA_CUSPARSE
615#define ALPHABETA(dtp, alpha, beta) \
616 __nv_bfloat16(alpha##16bf) = 1.0f; \
617 __nv_bfloat16(beta##16bf) = 1.0f; \
618 __half(alpha##16f) = 1.0f; \
619 __half(beta##16f) = 1.0f; \
620 float(alpha##f) = 1.0f; \
621 float(beta##f) = 1.0f; \
622 double(alpha##d) = 1.0; \
623 double(beta##d) = 1.0; \
624 const void *(alpha##p) = nullptr; \
625 const void *(beta##p) = nullptr; \
626 if (dtp == CUDA_R_16BF || dtp == CUDA_C_16BF) { \
627 (alpha##p) = reinterpret_cast<void *>(&(alpha##16bf)); \
628 (beta##p) = reinterpret_cast<void *>(&(beta##16bf)); \
629 } else if (dtp == CUDA_R_16F || dtp == CUDA_C_16F) { \
630 (alpha##p) = reinterpret_cast<void *>(&(alpha##16f)); \
631 (beta##p) = reinterpret_cast<void *>(&(beta##16f)); \
632 } else if (dtp == CUDA_R_32F || dtp == CUDA_C_32F) { \
633 (alpha##p) = reinterpret_cast<void *>(&(alpha##f)); \
634 (beta##p) = reinterpret_cast<void *>(&(beta##f)); \
636 (alpha##p) = reinterpret_cast<void *>(&(alpha##d)); \
637 (beta##p) = reinterpret_cast<void *>(&(beta##d)); \
643 assert(!cusparse_env &&
"client called mgpuCreateSparseEnv() twice");
648 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
650 cusparse_env =
nullptr;
654mgpuCreateDnVec(
intptr_t size,
void *values, int32_t dtp, CUstream ) {
655 cusparseDnVecDescr_t vec =
nullptr;
656 auto dTp =
static_cast<cudaDataType_t
>(dtp);
658 return reinterpret_cast<void *
>(vec);
662mgpuDestroyDnVec(
void *v, CUstream ) {
663 cusparseDnVecDescr_t vec =
reinterpret_cast<cusparseDnVecDescr_t
>(v);
670 cusparseDnMatDescr_t mat =
nullptr;
671 auto dTp =
static_cast<cudaDataType_t
>(dtp);
673 values, dTp, CUSPARSE_ORDER_ROW))
674 return reinterpret_cast<void *
>(mat);
678mgpuDestroyDnMat(
void *m, CUstream ) {
679 cusparseDnMatDescr_t mat =
reinterpret_cast<cusparseDnMatDescr_t
>(m);
685 void *colIdxs,
void *values, int32_t itp, int32_t dtp,
687 cusparseSpMatDescr_t mat =
nullptr;
688 auto iTp =
static_cast<cusparseIndexType_t
>(itp);
689 auto dTp =
static_cast<cudaDataType_t
>(dtp);
691 colIdxs, values, iTp,
692 CUSPARSE_INDEX_BASE_ZERO, dTp))
693 return reinterpret_cast<void *
>(mat);
696#ifdef CUSPARSE_COO_AOS
699 void *values, int32_t itp, int32_t dtp, CUstream ) {
700 cusparseSpMatDescr_t mat =
nullptr;
701 auto iTp =
static_cast<cusparseIndexType_t
>(itp);
702 auto dTp =
static_cast<cudaDataType_t
>(dtp);
704 &mat, rows, cols, nnz, idxs, values, iTp, CUSPARSE_INDEX_BASE_ZERO, dTp))
705 return reinterpret_cast<void *
>(mat);
711 void *colIdxs,
void *values, int32_t ptp, int32_t itp,
712 int32_t dtp, CUstream ) {
713 cusparseSpMatDescr_t mat =
nullptr;
714 auto pTp =
static_cast<cusparseIndexType_t
>(ptp);
715 auto iTp =
static_cast<cusparseIndexType_t
>(itp);
716 auto dTp =
static_cast<cudaDataType_t
>(dtp);
718 colIdxs, values, pTp, iTp,
719 CUSPARSE_INDEX_BASE_ZERO, dTp))
720 return reinterpret_cast<void *
>(mat);
725 void *rowIdxs,
void *values, int32_t ptp, int32_t itp,
726 int32_t dtp, CUstream ) {
727 cusparseSpMatDescr_t mat =
nullptr;
728 auto pTp =
static_cast<cusparseIndexType_t
>(ptp);
729 auto iTp =
static_cast<cusparseIndexType_t
>(itp);
730 auto dTp =
static_cast<cudaDataType_t
>(dtp);
732 rowIdxs, values, pTp, iTp,
733 CUSPARSE_INDEX_BASE_ZERO, dTp))
734 return reinterpret_cast<void *
>(mat);
739 intptr_t cBsz,
void *rowPos,
void *colIdxs,
void *values,
740 int32_t ptp, int32_t itp, int32_t dtp, CUstream ) {
741 cusparseSpMatDescr_t mat =
nullptr;
742#if CUSPARSE_VERSION >= 12100
743 auto pTp =
static_cast<cusparseIndexType_t
>(ptp);
744 auto iTp =
static_cast<cusparseIndexType_t
>(itp);
745 auto dTp =
static_cast<cudaDataType_t
>(dtp);
747 &mat, brows, bcols, bnnz, rBsz, cBsz, rowPos, colIdxs, values, pTp, iTp,
748 CUSPARSE_INDEX_BASE_ZERO, dTp, CUSPARSE_ORDER_ROW))
750 return reinterpret_cast<void *
>(mat);
754mgpuDestroySpMat(
void *m, CUstream ) {
755 cusparseSpMatDescr_t mat =
reinterpret_cast<cusparseSpMatDescr_t
>(m);
760 int32_t ma,
void *a,
void *x,
void *y, int32_t ctp, CUstream ) {
761 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
762 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
763 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
764 cusparseDnVecDescr_t vecX =
reinterpret_cast<cusparseDnVecDescr_t
>(x);
765 cusparseDnVecDescr_t vecY =
reinterpret_cast<cusparseDnVecDescr_t
>(y);
766 cudaDataType_t cTp =
static_cast<cudaDataType_t
>(ctp);
767 ALPHABETA(cTp, alpha, beta)
768 size_t bufferSize = 0;
770 cusparse_env, modeA, alphap, matA, vecX, betap, vecY, cTp,
771 CUSPARSE_SPMV_ALG_DEFAULT, &bufferSize))
776 void *y, int32_t ctp,
779 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
780 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
781 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
782 cusparseDnVecDescr_t vecX =
reinterpret_cast<cusparseDnVecDescr_t
>(x);
783 cusparseDnVecDescr_t vecY =
reinterpret_cast<cusparseDnVecDescr_t
>(y);
784 cudaDataType_t cTp =
static_cast<cudaDataType_t
>(ctp);
785 ALPHABETA(cTp, alpha, beta)
788 CUSPARSE_SPMV_ALG_DEFAULT, buf))
792mgpuSpMMBufferSize(int32_t ma, int32_t mb,
void *a,
void *
b,
void *c,
793 int32_t ctp, CUstream ) {
794 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
795 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
796 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
797 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
798 cusparseDnMatDescr_t matB =
reinterpret_cast<cusparseDnMatDescr_t
>(
b);
799 cusparseDnMatDescr_t matC =
reinterpret_cast<cusparseDnMatDescr_t
>(c);
800 cudaDataType_t cTp =
static_cast<cudaDataType_t
>(ctp);
801 ALPHABETA(cTp, alpha, beta)
802 size_t bufferSize = 0;
804 cusparse_env, modeA, modeB, alphap, matA, matB, betap, matC, cTp,
805 CUSPARSE_SPMM_ALG_DEFAULT, &bufferSize))
810 void *a,
void *
b,
void *c,
811 int32_t ctp,
void *buf,
813 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
814 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
815 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
816 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
817 cusparseDnMatDescr_t matB =
reinterpret_cast<cusparseDnMatDescr_t
>(
b);
818 cusparseDnMatDescr_t matC =
reinterpret_cast<cusparseDnMatDescr_t
>(c);
819 cudaDataType_t cTp =
static_cast<cudaDataType_t
>(ctp);
820 ALPHABETA(cTp, alpha, beta)
822 matA, matB, betap, matC, cTp,
823 CUSPARSE_SPMM_ALG_DEFAULT, buf))
827mgpuSDDMMBufferSize(int32_t ma, int32_t mb,
void *a,
void *
b,
void *c,
828 int32_t ctp, CUstream ) {
829 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
830 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
831 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
832 cusparseDnMatDescr_t matA =
reinterpret_cast<cusparseDnMatDescr_t
>(a);
833 cusparseDnMatDescr_t matB =
reinterpret_cast<cusparseDnMatDescr_t
>(
b);
834 cusparseSpMatDescr_t matC =
reinterpret_cast<cusparseSpMatDescr_t
>(c);
835 auto cTp =
static_cast<cudaDataType_t
>(ctp);
836 ALPHABETA(cTp, alpha, beta)
837 size_t bufferSize = 0;
839 cusparse_env, modeA, modeB, alphap, matA, matB, betap, matC, cTp,
840 CUSPARSE_SDDMM_ALG_DEFAULT, &bufferSize))
845 void *a,
void *
b,
void *c,
846 int32_t ctp,
void *buf,
848 assert(cusparse_env &&
"client did not call mgpuCreateSparseEnv()");
849 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
850 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
851 cusparseDnMatDescr_t matA =
reinterpret_cast<cusparseDnMatDescr_t
>(a);
852 cusparseDnMatDescr_t matB =
reinterpret_cast<cusparseDnMatDescr_t
>(
b);
853 cusparseSpMatDescr_t matC =
reinterpret_cast<cusparseSpMatDescr_t
>(c);
854 auto cTp =
static_cast<cudaDataType_t
>(ctp);
855 ALPHABETA(cTp, alpha, beta)
857 matA, matB, betap, matC, cTp,
858 CUSPARSE_SDDMM_ALG_DEFAULT, buf))
862mgpuSpGEMMCreateDescr(CUstream ) {
863 cusparseSpGEMMDescr_t spgemmDesc =
nullptr;
865 return reinterpret_cast<void *
>(spgemmDesc);
869mgpuSpGEMMDestroyDescr(
void *s, CUstream ) {
870 cusparseSpGEMMDescr_t spgemmDesc =
reinterpret_cast<cusparseSpGEMMDescr_t
>(s);
875 void *s, int32_t ma, int32_t mb,
void *a,
void *
b,
void *c, int32_t ctp,
876 intptr_t bs,
void *buf, CUstream ) {
877 cusparseSpGEMMDescr_t spgemmDesc =
reinterpret_cast<cusparseSpGEMMDescr_t
>(s);
878 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
879 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
880 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
881 cusparseSpMatDescr_t matB =
reinterpret_cast<cusparseSpMatDescr_t
>(
b);
882 cusparseSpMatDescr_t matC =
reinterpret_cast<cusparseSpMatDescr_t
>(c);
883 auto cTp =
static_cast<cudaDataType_t
>(ctp);
884 ALPHABETA(cTp, alpha, beta)
885 size_t newBufferSize = bs;
887 cusparse_env, modeA, modeB, alphap, matA, matB, betap, matC, cTp,
888 CUSPARSE_SPGEMM_DEFAULT, spgemmDesc, &newBufferSize, buf))
889 return newBufferSize;
893mgpuSpGEMMCompute(
void *s, int32_t ma, int32_t mb,
void *a,
void *
b,
void *c,
894 int32_t ctp,
intptr_t bsz2,
void *buf2, CUstream ) {
895 cusparseSpGEMMDescr_t spgemmDesc =
reinterpret_cast<cusparseSpGEMMDescr_t
>(s);
896 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
897 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
898 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
899 cusparseSpMatDescr_t matB =
reinterpret_cast<cusparseSpMatDescr_t
>(
b);
900 cusparseSpMatDescr_t matC =
reinterpret_cast<cusparseSpMatDescr_t
>(c);
901 auto cTp =
static_cast<cudaDataType_t
>(ctp);
902 ALPHABETA(cTp, alpha, beta)
903 size_t newBufferSize2 = bsz2;
905 cusparse_env, modeA, modeB, alphap, matA, matB, betap, matC, cTp,
906 CUSPARSE_SPGEMM_DEFAULT, spgemmDesc, &newBufferSize2, buf2))
907 return newBufferSize2;
911mgpuSpGEMMCopy(
void *s, int32_t ma, int32_t mb,
void *a,
void *
b,
void *c,
912 int32_t ctp, CUstream ) {
913 cusparseSpGEMMDescr_t spgemmDesc =
reinterpret_cast<cusparseSpGEMMDescr_t
>(s);
914 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
915 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
916 cusparseSpMatDescr_t matA =
reinterpret_cast<cusparseSpMatDescr_t
>(a);
917 cusparseSpMatDescr_t matB =
reinterpret_cast<cusparseSpMatDescr_t
>(
b);
918 cusparseSpMatDescr_t matC =
reinterpret_cast<cusparseSpMatDescr_t
>(c);
919 auto cTp =
static_cast<cudaDataType_t
>(ctp);
920 ALPHABETA(cTp, alpha, beta)
922 cusparseSpGEMM_copy(cusparse_env, modeA, modeB, alphap, matA, matB, betap,
923 matC, cTp, CUSPARSE_SPGEMM_DEFAULT, spgemmDesc))
927mgpuSpMatGetSize(
void *m,
void *r,
void *c,
void *n, CUstream ) {
928 cusparseConstSpMatDescr_t matDescr =
929 reinterpret_cast<cusparseConstSpMatDescr_t
>(m);
937mgpuSetCsrPointers(
void *m,
void *p,
void *c,
void *v, CUstream ) {
938 cusparseSpMatDescr_t matDescr =
reinterpret_cast<cusparseSpMatDescr_t
>(m);
942#ifdef MLIR_ENABLE_CUDA_CUSPARSELT
948struct cusparseLtSpMatHandleAndData {
949 cusparseLtMatDescriptor_t mat;
953 cusparseLtMatmulAlgSelection_t alg_sel;
954 cusparseLtMatmulPlan_t plan;
955 cusparseLtMatmulDescriptor_t matmul;
956 void *values{
nullptr};
959struct cusparseLtDnMatHandleAndData {
960 cusparseLtMatDescriptor_t mat;
961 void *values{
nullptr};
964static_assert(
sizeof(cusparseLtHandle_t) == 11024,
965 "Unexpected cusparseLt handle size");
966static_assert(
sizeof(cusparseLtSpMatHandleAndData) == 44104,
967 "Unexpected cusparseLt sparse matrix handle size");
968static_assert(
sizeof(cusparseLtDnMatHandleAndData) == 11032,
969 "Unexpected cusparseLt dense matrix handle size");
974 assert(!cusparseLt_initiated &&
975 "client called mgpuCreateSparseLtEnv() twice");
978 cusparseLt_initiated =
true;
982 assert(cusparseLt_initiated &&
"client did not call mgpuCreateSparseLtEnv()");
984 cusparseLt_initiated =
false;
988mgpuCreateCuSparseLtDnMat(
void *dh,
intptr_t rows,
intptr_t cols,
void *values,
989 int32_t dtp, CUstream ) {
990 assert(cusparseLt_initiated &&
"client did not call mgpuCreateSparseLtEnv()");
991 auto dnmat_handle =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(dh);
992 dnmat_handle->values = values;
993 auto dTp =
static_cast<cudaDataType_t
>(dtp);
995 const uint32_t alignment = 16;
997 &cusparseLt_env, &(dnmat_handle->mat), rows, cols, cols,
998 alignment, dTp, CUSPARSE_ORDER_ROW))
1002mgpuDestroyCuSparseLtDnMat(
void *dh, CUstream ) {
1003 auto dnmat_handle =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(dh);
1009 void *values, int32_t dtp, CUstream ) {
1010 assert(cusparseLt_initiated &&
"client did not call mgpuCreateSparseLtEnv()");
1011 auto spmat_handle =
reinterpret_cast<cusparseLtSpMatHandleAndData *
>(sh);
1012 spmat_handle->values = values;
1013 auto dTp =
static_cast<cudaDataType_t
>(dtp);
1015 const uint32_t alignment = 16;
1017 &cusparseLt_env, &(spmat_handle->mat), rows, cols, cols, alignment,
1018 dTp, CUSPARSE_ORDER_ROW, CUSPARSELT_SPARSITY_50_PERCENT))
1022mgpuDestroyCuSparseLtSpMat(
void *sh, CUstream ) {
1023 auto spmat_handle =
reinterpret_cast<cusparseLtSpMatHandleAndData *
>(sh);
1033mgpuCuSparseLtSpMMBufferSize(
void *bs, int32_t ma, int32_t mb,
void *a,
void *
b,
1034 void *c, int32_t ctp, int32_t prune_flag,
1036 assert(cusparseLt_initiated &&
"client did not call mgpuCreateSparseLtEnv()");
1039 auto matA =
reinterpret_cast<cusparseLtSpMatHandleAndData *
>(a);
1040 auto matB =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(
b);
1041 auto matC =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(c);
1042 auto workspace_size =
reinterpret_cast<size_t *
>(bs);
1043 auto compressed_size = &(
reinterpret_cast<size_t *
>(bs)[1]);
1044 auto compressed_buffer_size = &(
reinterpret_cast<size_t *
>(bs)[2]);
1045 auto cTp =
static_cast<cusparseComputeType
>(ctp);
1047 cusparseOperation_t modeA =
static_cast<cusparseOperation_t
>(ma);
1048 cusparseOperation_t modeB =
static_cast<cusparseOperation_t
>(mb);
1050 &cusparseLt_env, &(matA->matmul), modeA, modeB, &(matA->mat),
1051 &(matB->mat), &(matC->mat), &(matC->mat), cTp))
1053 &cusparseLt_env, &(matA->alg_sel), &(matA->matmul),
1054 CUSPARSELT_MATMUL_ALG_DEFAULT))
1057 &cusparseLt_env, &(matA->alg_sel), CUSPARSELT_MATMUL_ALG_CONFIG_ID, &alg,
1061 &cusparseLt_env, &(matA->plan), &(matA->matmul), &(matA->alg_sel)))
1066 &cusparseLt_env, &(matA->matmul), matA->values, matA->values,
1067 CUSPARSELT_PRUNE_SPMMA_STRIP, stream))
1072 if (prune_flag == 2) {
1073 int *dvalid = (
int *)
mgpuMemAlloc(
sizeof(
int), stream,
false);
1075 &cusparseLt_env, &(matA->matmul), matA->values, dvalid, stream))
1077 mgpuMemcpy(&valid, dvalid,
sizeof(
int), stream);
1081 fprintf(stderr,
"CUPARSE-LT: sparse matrix is not 2:4; computed results "
1082 "will be invalid\n");
1086 &cusparseLt_env, &(matA->plan), workspace_size))
1088 &cusparseLt_env, &(matA->plan), compressed_size, compressed_buffer_size))
1092mgpuCuSparseLtSpMM(
void *a,
void *
b,
void *c,
void *d_workspace,
1093 void *dA_compressed,
void *dA_compressedBuffer,
1095 assert(cusparseLt_initiated &&
"client did not call mgpuCreateSparseLtEnv()");
1096 auto matA =
reinterpret_cast<cusparseLtSpMatHandleAndData *
>(a);
1097 auto matB =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(
b);
1098 auto matC =
reinterpret_cast<cusparseLtDnMatHandleAndData *
>(c);
1100 ALPHABETA(CUDA_R_32F, alpha, beta)
1102 cusparseLtSpMMACompress(&cusparseLt_env, &(matA->plan), (matA->values),
1103 dA_compressed, dA_compressedBuffer, stream))
1108 cusparseLtMatmul(&cusparseLt_env, &(matA->plan), alphap, dA_compressed,
1109 matB->values, betap, matC->values,
1110 matC->values, d_workspace,
nullptr, 0))
#define CUSPARSE_REPORT_IF_ERROR(expr)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuStreamWaitEvent(CUstream stream, CUevent event)
#define MLIR_CUDA_WRAPPERS_EXPORT
MLIR_CUDA_WRAPPERS_EXPORT void mgpuModuleUnload(CUmodule module)
MLIR_CUDA_WRAPPERS_EXPORT void * mgpuMemAlloc(uint64_t sizeBytes, CUstream stream, bool isHostShared)
#define CUDA_REPORT_IF_ERROR(expr)
static bool isCudaContextShutdownError(CUresult result)
Helper to check if a CUDA error is due to the context being destroyed during program shutdown.
#define CUDA_REPORT_IF_ERROR_IGNORE_SHUTDOWN(expr)
Like CUDA_REPORT_IF_ERROR, but silences errors caused by CUDA context shutdown.
static thread_local int32_t defaultDevice
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemHostRegisterMemRef(int64_t rank, StridedMemRefType< char, 1 > *descriptor, int64_t elementSizeBytes)
Registers a memref with the CUDA runtime.
MLIR_CUDA_WRAPPERS_EXPORT CUmodule mgpuModuleLoadJIT(void *data, int optLevel, size_t)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemHostUnregister(void *ptr)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuEventRecord(CUevent event, CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemcpy(void *dst, void *src, size_t sizeBytes, CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemset16(void *dst, unsigned short value, size_t count, CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemset32(void *dst, unsigned int value, size_t count, CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT CUfunction mgpuModuleGetFunction(CUmodule module, const char *name)
#define debug_print(fmt,...)
MLIR_CUDA_WRAPPERS_EXPORT CUevent mgpuEventCreate()
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemHostRegister(void *ptr, uint64_t sizeBytes)
Helper functions for writing mlir example code.
static bool isDebugEnabled()
Helper method that checks environment value for debugging.
MLIR_CUDA_WRAPPERS_EXPORT void mgpuStreamSynchronize(CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT CUstream mgpuStreamCreate()
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemFree(void *ptr, CUstream)
MLIR_CUDA_WRAPPERS_EXPORT CUmodule mgpuModuleLoad(void *data, size_t)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuSetDefaultDevice(int32_t device)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuStreamDestroy(CUstream stream)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuLaunchKernel(CUfunction function, intptr_t gridX, intptr_t gridY, intptr_t gridZ, intptr_t blockX, intptr_t blockY, intptr_t blockZ, int32_t smem, CUstream stream, void **params, void **extra, size_t)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuMemHostUnregisterMemRef(int64_t rank, StridedMemRefType< char, 1 > *descriptor, int64_t elementSizeBytes)
Unregisters a memref with the CUDA runtime.
MLIR_CUDA_WRAPPERS_EXPORT void mgpuEventSynchronize(CUevent event)
MLIR_CUDA_WRAPPERS_EXPORT void mgpuEventDestroy(CUevent event)
static CUdevice getDefaultCuDevice()
void mgpuLaunchKernelCooperative(hipFunction_t function, intptr_t gridX, intptr_t gridY, intptr_t gridZ, intptr_t clusterX, intptr_t clusterY, intptr_t clusterZ, intptr_t blockX, intptr_t blockY, intptr_t blockZ, int32_t smem, hipStream_t stream, void **params, void **)
StridedMemRef descriptor type with static rank.