From 6b3058b2dc9d45f38e79b293604e6391ae88b48d Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Fri, 6 Dec 2024 09:08:55 -0500 Subject: [PATCH] upgrade bnb 0.45.0 and peft 0.14.0 (#2126) * upgrade bnb to lastest release * update peft to working supporting commit * bump to latest release of peft==0.14.0 --- requirements.txt | 4 ++-- .../utils/config/models/input/v0_4_1/__init__.py | 13 ------------- tests/e2e/multigpu/test_llama.py | 3 --- 3 files changed, 2 insertions(+), 18 deletions(-) diff --git a/requirements.txt b/requirements.txt index 456c63ca51..d100139ca4 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,9 +1,9 @@ --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/ packaging==23.2 -peft==0.13.2 +peft==0.14.0 transformers==4.46.3 tokenizers>=0.20.1 -bitsandbytes==0.44.1 +bitsandbytes==0.45.0 accelerate==1.1.0 datasets==3.1.0 deepspeed==0.15.4 diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py index c9170b7a84..24ea62c77f 100644 --- a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py +++ b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py @@ -1521,19 +1521,6 @@ def check_sample_packing_w_sdpa_bf16(cls, data): return data - @model_validator(mode="before") - @classmethod - def check_hopper_8bit_lora(cls, data): - is_sm_90: bool = ( - data["capabilities"] - and data["capabilities"].get("compute_capability") == "sm_90" - ) - if data.get("adapter") and data.get("load_in_8bit") and is_sm_90: - # see https://github.com/bitsandbytes-foundation/bitsandbytes/issues/538#issuecomment-2262945464 - raise ValueError("8-bit LoRA is not supported on Hopper GPUs") - - return data - @model_validator(mode="before") @classmethod def check_fsdp_deepspeed(cls, data): diff --git a/tests/e2e/multigpu/test_llama.py b/tests/e2e/multigpu/test_llama.py index d8dcf3118a..c9938a1559 100644 --- a/tests/e2e/multigpu/test_llama.py +++ b/tests/e2e/multigpu/test_llama.py @@ -14,8 +14,6 @@ from axolotl.utils.dict import DictDefault -from ..utils import is_hopper - LOG = logging.getLogger("axolotl.tests.e2e.multigpu") os.environ["WANDB_DISABLED"] = "true" @@ -144,7 +142,6 @@ def test_lora_ddp_packed(self, temp_dir, gradient_accumulation_steps): ] ) - @pytest.mark.skipif(is_hopper(), reason="h100 doesn't support 8-bit lora") def test_dpo_lora_ddp(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault(