warp submodule#
- class cutlass.cute.nvgpu.warp.MmaF16BF16Op(
- ab_dtype: Type[cutlass.cute.typing.Numeric],
- acc_dtype: Type[cutlass.cute.typing.Numeric],
- shape_mnk: cutlass.cute.typing.Shape,
Bases:
MmaOp
F16/BF16 tcgen05 MMA Operation.
See the PTX documentation. This Operation covers the instructions using the
.f16
or.bf16
qualifiers for the input operands.- ab_dtype: Type[cutlass.cute.typing.Numeric]#
- acc_dtype: Type[cutlass.cute.typing.Numeric]#
- shape_mnk: cutlass.cute.typing.Shape#
- __init__(
- ab_dtype: Type[cutlass.cute.typing.Numeric],
- acc_dtype: Type[cutlass.cute.typing.Numeric],
- shape_mnk: cutlass.cute.typing.Shape,
- class cutlass.cute.nvgpu.warp.LdMatrix8x8x16bOp(transpose: bool = False, num_matrices: int = 1)#
Bases:
BaseOp
8x8
ldmatrix
Operation.See the PTX documentation. This operation corresponds to the
.m8n8
qualifier.- __init__(
- transpose: bool = False,
- num_matrices: int = 1,
- class cutlass.cute.nvgpu.warp.LdMatrix16x16x8bOp(num_matrices: int)#
Bases:
BaseOp
16x16 8-bit
ldmatrix
Operation.See the PTX documentation. This operation corresponds to the
.m16n16
and the.b16
qualifiers.- __init__(num_matrices: int) None #
- class cutlass.cute.nvgpu.warp.StMatrix8x8x16bOp(transpose: bool = False, num_matrices: int = 1)#
Bases:
BaseOp
8x8
stmatrix
Operation.See the PTX documentation. This operation corresponds to the
m8n8
qualifier.- __init__(
- transpose: bool = False,
- num_matrices: int = 1,
- class cutlass.cute.nvgpu.warp.StMatrix16x8x8bOp(num_matrices: int)#
Bases:
BaseOp
16x8
stmatrix
Operation.See the PTX documentation. This operation corresponds to the
m16n8
qualifier.- __init__(num_matrices: int) None #