There was an error while loading. Please reload this page.
1 parent f3a8e8d commit 2895626Copy full SHA for 2895626
test/dtypes/test_affine_quantized_float.py
@@ -288,8 +288,10 @@ def test_fp8_weight_dimension_warning(self):
288
)
289
290
291
-@unittest.skip("Only running locally so we don't need to add installation of fa3 "
292
- "hopper kernels to CI, we'll probably copy paste kernel in the future")
+@unittest.skip(
+ "Only running locally so we don't need to add installation of fa3 "
293
+ "hopper kernels to CI, we'll probably copy paste kernel in the future"
294
+)
295
class TestAffineQuantizedFloat8Attention(common_utils.TestCase):
296
@unittest.skipIf(not torch.cuda.is_available(), "Need CUDA available")
297
def test_float8_attention(self):
0 commit comments