Commit d848d4db authored by Matthew Douglas's avatar Matthew Douglas
Browse files

Lint fix

parent 6a07ffe0
...@@ -1418,7 +1418,11 @@ class TestQuantize4BitFunctional: ...@@ -1418,7 +1418,11 @@ class TestQuantize4BitFunctional:
if device == "hpu" and not is_supported_on_hpu(storage_type, dtype): if device == "hpu" and not is_supported_on_hpu(storage_type, dtype):
pytest.skip("This configuration is not supported on HPU.") pytest.skip("This configuration is not supported on HPU.")
if device == "cpu" and platform.system() == "Windows" and version.parse(torch.__version__).release == (2, 8, 0): if (
device == "cpu"
and platform.system() == "Windows"
and version.parse(torch.__version__).release == (2, 8, 0)
):
pytest.skip("Regression: CPU crash on Windows with torch 2.8.0") pytest.skip("Regression: CPU crash on Windows with torch 2.8.0")
dims = 4 dims = 4
......
...@@ -172,7 +172,6 @@ optimizer_names_32bit = [ ...@@ -172,7 +172,6 @@ optimizer_names_32bit = [
@pytest.mark.parametrize("device", get_available_devices(no_cpu=True), ids=id_formatter("device")) @pytest.mark.parametrize("device", get_available_devices(no_cpu=True), ids=id_formatter("device"))
@pytest.mark.skipif(not get_available_devices(no_cpu=True), reason="No device") @pytest.mark.skipif(not get_available_devices(no_cpu=True), reason="No device")
def test_optimizer32bit(dim1, dim2, gtype, optim_name, device): def test_optimizer32bit(dim1, dim2, gtype, optim_name, device):
if device not in ["cuda", "xpu"]: if device not in ["cuda", "xpu"]:
pytest.skip("Optimizers are only supported on CUDA and XPU") pytest.skip("Optimizers are only supported on CUDA and XPU")
...@@ -317,7 +316,6 @@ optimizer_names_8bit = [ ...@@ -317,7 +316,6 @@ optimizer_names_8bit = [
@pytest.mark.parametrize("device", get_available_devices(no_cpu=True)) @pytest.mark.parametrize("device", get_available_devices(no_cpu=True))
@pytest.mark.skipif(not get_available_devices(no_cpu=True), reason="No device") @pytest.mark.skipif(not get_available_devices(no_cpu=True), reason="No device")
def test_optimizer8bit(dim1, dim2, gtype, optim_name, device): def test_optimizer8bit(dim1, dim2, gtype, optim_name, device):
if device not in ["cuda", "xpu"]: if device not in ["cuda", "xpu"]:
pytest.skip("8-bit optimizers are only supported on CUDA and XPU") pytest.skip("8-bit optimizers are only supported on CUDA and XPU")
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment