@@ -74,28 +74,6 @@ def matmul_mxf4_bf16_tn(
7474 raise ValueError (f"invalid backend { backend !r} ; use 'cutlass' or 'flashinfer'" )
7575
7676
77- def matmul_mxf8_mxf4_bf16_tn (a : torch .Tensor ,
78- b : torch .Tensor ,
79- a_sf : torch .Tensor ,
80- b_sf : torch .Tensor ,
81- alpha : torch .Tensor ) -> torch .Tensor :
82- return qutlass ._CUDA .matmul_mxf8_mxf4_bf16_tn (a , b , a_sf , b_sf , alpha )
83-
84- def matmul_mxf8_mxf4_bf16_nt (a : torch .Tensor ,
85- b : torch .Tensor ,
86- a_sf : torch .Tensor ,
87- b_sf : torch .Tensor ,
88- alpha : torch .Tensor ) -> torch .Tensor :
89- return qutlass ._CUDA .matmul_mxf8_mxf4_bf16_nt (a , b , a_sf , b_sf , alpha )
90-
91- def matmul_mxf8_mxf4_bf16_tt (a : torch .Tensor ,
92- b : torch .Tensor ,
93- a_sf : torch .Tensor ,
94- b_sf : torch .Tensor ,
95- alpha : torch .Tensor ) -> torch .Tensor :
96- return qutlass ._CUDA .matmul_mxf8_mxf4_bf16_tt (a , b , a_sf , b_sf , alpha )
97-
98-
9977def matmul_ada_mxf4_bf16_tn (
10078 a : torch .Tensor ,
10179 b : torch .Tensor ,
@@ -158,20 +136,6 @@ def matmul_mxf8_bf16_tn(a: torch.Tensor,
158136 alpha : torch .Tensor ) -> torch .Tensor :
159137 return qutlass ._CUDA .matmul_mxf8_bf16_tn (a , b , block_scale_a , block_scale_b , alpha )
160138
161- def matmul_mxf8_bf16_nt (a : torch .Tensor ,
162- b : torch .Tensor ,
163- block_scale_a : torch .Tensor ,
164- block_scale_b : torch .Tensor ,
165- alpha : torch .Tensor ) -> torch .Tensor :
166- return qutlass ._CUDA .matmul_mxf8_bf16_nt (a , b , block_scale_a , block_scale_b , alpha )
167-
168- def matmul_mxf8_bf16_tt (a : torch .Tensor ,
169- b : torch .Tensor ,
170- block_scale_a : torch .Tensor ,
171- block_scale_b : torch .Tensor ,
172- alpha : torch .Tensor ) -> torch .Tensor :
173- return qutlass ._CUDA .matmul_mxf8_bf16_tt (a , b , block_scale_a , block_scale_b , alpha )
174-
175139def matmul_mxf8_bf16_nn (a : torch .Tensor ,
176140 b : torch .Tensor ,
177141 block_scale_a : torch .Tensor ,
0 commit comments