We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 6230a5d commit 54e7384Copy full SHA for 54e7384
tests/models/quantization/test_bitsandbytes.py
@@ -19,7 +19,7 @@
19
20
pytestmark = pytest.mark.skipif(
21
on_gfx9(),
22
- reason="bitsandbytes quantization not supported on gfx9 (warp size 64 limitation)",
+ reason="bitsandbytes quantization not supported on gfx9 (warp size 64)",
23
)
24
25
models_4bit_to_test = [
vllm/platforms/rocm.py
@@ -185,7 +185,7 @@ class RocmPlatform(Platform):
185
"petit_nvfp4",
186
"torchao",
187
]
188
- # bitsandbytes quantization not supported on gfx9 (warp size 64 limitation)
+ # bitsandbytes not supported on gfx9 (warp size 64 limitation)
189
if not on_gfx9():
190
supported_quantization += ["bitsandbytes"]
191
0 commit comments