Skip to content

Commit

Permalink
Fix for the current state of platform specific quantizations
Browse files Browse the repository at this point in the history
  • Loading branch information
gshtras committed Dec 2, 2024
1 parent 1e93ebe commit a8b5334
Showing 1 changed file with 5 additions and 10 deletions.
15 changes: 5 additions & 10 deletions vllm/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -443,16 +443,11 @@ def _verify_quantization(self) -> None:
method = get_quantization_config(name)
quantization_override = method.override_quantization_method(
quant_cfg, self.quantization)
if quantization_override:
if current_platform.is_rocm():
if quantization_override in rocm_supported_quantization:
quant_method = quantization_override
self.quantization = quantization_override
break
else:
quant_method = quantization_override
self.quantization = quantization_override
break
if (quantization_override and quantization_override
in current_platform.supported_quantization):
quant_method = quantization_override
self.quantization = quantization_override
break

# Verify quantization configurations.
if self.quantization is None:
Expand Down

0 comments on commit a8b5334

Please sign in to comment.