warp submodule#
- class cutlass.cute.nvgpu.warp.MmaF16BF16Op(
- ab_dtype: Type[cutlass.cute.typing.Numeric],
- acc_dtype: Type[cutlass.cute.typing.Numeric],
- shape_mnk: cutlass.cute.typing.Shape,
Bases:
WarpMmaOpF16/BF16 tcgen05 MMA Operation.
See the PTX documentation. This Operation covers the instructions using the
.f16or.bf16qualifiers for the input operands.- ab_dtype: Type[cutlass.cute.typing.Numeric]#
- acc_dtype: Type[cutlass.cute.typing.Numeric]#
- shape_mnk: cutlass.cute.typing.Shape#
- __init__(
- ab_dtype: Type[cutlass.cute.typing.Numeric],
- acc_dtype: Type[cutlass.cute.typing.Numeric],
- shape_mnk: cutlass.cute.typing.Shape,
- class cutlass.cute.nvgpu.warp.LdMatrix8x8x16bOp(transpose: bool = False, num_matrices: int = 1)#
Bases:
BaseOp8x8
ldmatrixOperation.See the PTX documentation. This operation corresponds to the
.m8n8qualifier.- __init__(
- transpose: bool = False,
- num_matrices: int = 1,
- class cutlass.cute.nvgpu.warp.LdMatrix16x16x8bOp(num_matrices: int)#
Bases:
BaseOp16x16 8-bit
ldmatrixOperation.See the PTX documentation. This operation corresponds to the
.m16n16and the.b16qualifiers.- __init__(num_matrices: int) None#
- class cutlass.cute.nvgpu.warp.StMatrix8x8x16bOp(transpose: bool = False, num_matrices: int = 1)#
Bases:
BaseOp8x8
stmatrixOperation.See the PTX documentation. This operation corresponds to the
m8n8qualifier.- __init__(
- transpose: bool = False,
- num_matrices: int = 1,
- class cutlass.cute.nvgpu.warp.StMatrix16x8x8bOp(num_matrices: int)#
Bases:
BaseOp16x8
stmatrixOperation.See the PTX documentation. This operation corresponds to the
m16n8qualifier.- __init__(num_matrices: int) None#