From ce6b0e320c7b1e26f0086384529a1b356319b88e Mon Sep 17 00:00:00 2001 From: Giuseppe Franco Date: Thu, 22 Aug 2024 14:16:39 +0100 Subject: [PATCH] Fix tests --- tests/brevitas/nn/nn_quantizers_fixture.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/tests/brevitas/nn/nn_quantizers_fixture.py b/tests/brevitas/nn/nn_quantizers_fixture.py index c3102153f..73258bd9a 100644 --- a/tests/brevitas/nn/nn_quantizers_fixture.py +++ b/tests/brevitas/nn/nn_quantizers_fixture.py @@ -127,12 +127,10 @@ def build_case_model( pytest.skip( "A2Q uses an input-aware decoupled weight proxy that requires a quantized input tensor." ) - if weight_quantizer == ShiftedUint8WeightPerGroupFloatHQO and bias_quantizer == Int16Bias: - pytest.skip("HQO doesn't allow for bias with external scale") - if weight_quantizer == ShiftedUint8WeightPerGroupFloatHQO and bias_quantizer == Int16Bias: + if weight_quantizer == TestHQO and bias_quantizer == Int16Bias: pytest.skip("HQO doesn't allow for bias with external scale") impl = module.__name__ - if weight_quantizer == ShiftedUint8WeightPerGroupFloatHQO and 'Transpose' in impl: + if weight_quantizer == TestHQO and 'Transpose' in impl: pytest.skip("HQO doesn't support transposed convolution") # BatchQuant has dimension specific quantizers if isinstance(io_quantizer, tuple):