Skip to content

Commit

Permalink
[bugfix] fix the default value of llm_int8_threshold in BitsAndBytesC…
Browse files Browse the repository at this point in the history
…onfig (vllm-project#10657)

Signed-off-by: Andrew Feldman <[email protected]>
  • Loading branch information
yansh97 authored and afeldman-nm committed Dec 2, 2024
1 parent 0f196ac commit 429d17e
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions vllm/model_executor/layers/quantization/bitsandbytes.py
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ def __init__(
llm_int8_enable_fp32_cpu_offload: bool = False,
llm_int8_has_fp16_weight: bool = False,
llm_int8_skip_modules: Optional[List[str]] = None,
llm_int8_threshold: float = 0.0,
llm_int8_threshold: float = 6.0,
) -> None:

self.load_in_8bit = load_in_8bit
Expand Down Expand Up @@ -103,7 +103,7 @@ def get_safe_value(config, keys, default_value=None):
["llm_int8_skip_modules"],
default_value=[])
llm_int8_threshold = get_safe_value(config, ["llm_int8_threshold"],
default_value=0.0)
default_value=6.0)

return cls(
load_in_8bit=load_in_8bit,
Expand Down

0 comments on commit 429d17e

Please sign in to comment.