mlir.dialects._gpu_enum_gen¶
Attributes¶
Classes¶
GPU address space |
|
built-in reduction operations supported by gpu.allreduce. |
|
a lane to broadcast from |
|
GPU compilation format |
|
a dimension, either 'x', 'y', or 'z' |
|
pruning strategy for 2:4 sparse matrix |
|
Indexing modes supported by gpu.shuffle. |
|
choose whether spgemm_work_estimation_or_compute does work estimation or compute |
|
transpose mode of sparse matrix supported by sparse tensor ops |
|
elementwise operation to apply to mma matrix |
|
Mapping ids for loop mapping |
|
processor for loop mapping |
Functions¶
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Module Contents¶
- mlir.dialects._gpu_enum_gen._ods_ir¶
- class mlir.dialects._gpu_enum_gen.AddressSpace¶
Bases:
enum.IntEnumGPU address space
- Global = 1¶
- Workgroup = 2¶
- Private = 3¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_addressspaceenum(x, context)¶
- class mlir.dialects._gpu_enum_gen.AllReduceOperation¶
Bases:
enum.IntEnumbuilt-in reduction operations supported by gpu.allreduce.
- ADD = 0¶
- MUL = 1¶
- MINUI = 2¶
- MINSI = 3¶
- MINNUMF = 4¶
- MAXUI = 5¶
- MAXSI = 6¶
- MAXNUMF = 7¶
- AND = 8¶
- OR = 9¶
- XOR = 10¶
- MINIMUMF = 11¶
- MAXIMUMF = 12¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_allreduceoperation(x, context)¶
- class mlir.dialects._gpu_enum_gen.BroadcastType¶
Bases:
enum.IntEnuma lane to broadcast from
- first_active_lane = 0¶
- specific_lane = 1¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_broadcasttype(x, context)¶
- class mlir.dialects._gpu_enum_gen.CompilationTarget¶
Bases:
enum.IntEnumGPU compilation format
- Offload = 1¶
- Assembly = 2¶
- Binary = 3¶
- Fatbin = 4¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_compilationtargetenum(x, context)¶
- class mlir.dialects._gpu_enum_gen.Dimension¶
Bases:
enum.IntEnuma dimension, either ‘x’, ‘y’, or ‘z’
- x = 0¶
- y = 1¶
- z = 2¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_dimension(x, context)¶
- class mlir.dialects._gpu_enum_gen.Prune2To4SpMatFlag¶
Bases:
enum.IntEnumpruning strategy for 2:4 sparse matrix
- NONE = 0¶
- PRUNE_ONLY = 1¶
- PRUNE_AND_CHECK = 2¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_prune2to4spmatflag(x, context)¶
- class mlir.dialects._gpu_enum_gen.ShuffleMode¶
Bases:
enum.IntEnumIndexing modes supported by gpu.shuffle.
- XOR = 0¶
- UP = 2¶
- DOWN = 1¶
- IDX = 3¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_shufflemode(x, context)¶
- class mlir.dialects._gpu_enum_gen.SpGEMMWorkEstimationOrComputeKind¶
Bases:
enum.IntEnumchoose whether spgemm_work_estimation_or_compute does work estimation or compute
- WORK_ESTIMATION = 0¶
- COMPUTE = 1¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_spgemmworkestimationorcomputekind(x, context)¶
- class mlir.dialects._gpu_enum_gen.TransposeMode¶
Bases:
enum.IntEnumtranspose mode of sparse matrix supported by sparse tensor ops
- NON_TRANSPOSE = 0¶
- TRANSPOSE = 1¶
- CONJUGATE_TRANSPOSE = 2¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._gpu_transposemode(x, context)¶
- class mlir.dialects._gpu_enum_gen.MMAElementwiseOp¶
Bases:
enum.IntEnumelementwise operation to apply to mma matrix
- ADDF = 0¶
- MULF = 1¶
- SUBF = 2¶
- MAXF = 3¶
- MINF = 4¶
- DIVF = 5¶
- ADDI = 6¶
- MULI = 7¶
- SUBI = 8¶
- DIVS = 9¶
- DIVU = 10¶
- NEGATEF = 11¶
- NEGATES = 12¶
- EXTF = 13¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._mmaelementwise(x, context)¶
- class mlir.dialects._gpu_enum_gen.MappingId¶
Bases:
enum.IntEnumMapping ids for loop mapping
- DimX = 0¶
- DimY = 1¶
- DimZ = 2¶
- LinearDim0 = 3¶
- LinearDim1 = 4¶
- LinearDim2 = 5¶
- LinearDim3 = 6¶
- LinearDim4 = 7¶
- LinearDim5 = 8¶
- LinearDim6 = 9¶
- LinearDim7 = 10¶
- LinearDim8 = 11¶
- LinearDim9 = 12¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._mappingidenum(x, context)¶
- class mlir.dialects._gpu_enum_gen.Processor¶
Bases:
enum.IntEnumprocessor for loop mapping
- BlockX = 0¶
- BlockY = 1¶
- BlockZ = 2¶
- ThreadX = 3¶
- ThreadY = 4¶
- ThreadZ = 5¶
- Sequential = 6¶
- __str__()¶
Return str(self).
- mlir.dialects._gpu_enum_gen._processorenum(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_addressspaceattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_allreduceoperationattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_broadcasttypeattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_dimensionattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_prune2to4spmatflagattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_shufflemodeattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_spgemmworkestimationorcomputekindattr(x, context)¶
- mlir.dialects._gpu_enum_gen._gpu_transposemodeattr(x, context)¶
- mlir.dialects._gpu_enum_gen._mmaelementwiseattr(x, context)¶