forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Showing
5 changed files
with
103 additions
and
3 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Empty file.
32 changes: 32 additions & 0 deletions
32
tests/runai_model_streamer/test_runai_model_streamer_loader.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,32 @@ | ||
from vllm.model_executor.model_loader.loader import (get_model_loader, RunaiModelStreamerLoader) | ||
from vllm.config import (LoadConfig, LoadFormat) | ||
from ..conftest import VllmRunner | ||
from vllm import SamplingParams | ||
|
||
|
||
test_model = "openai-community/gpt2" | ||
|
||
prompts = [ | ||
"Hello, my name is", | ||
"The president of the United States is", | ||
"The capital of France is", | ||
"The future of AI is", | ||
] | ||
# Create a sampling params object. | ||
sampling_params = SamplingParams(temperature=0.8, top_p=0.95, seed=0) | ||
|
||
def get_runai_model_loader(): | ||
load_config = LoadConfig(load_format=LoadFormat.RUNAI_STREAMER) | ||
return get_model_loader(load_config) | ||
|
||
def test_get_model_loader_with_runai_flag(): | ||
model_loader = get_runai_model_loader() | ||
assert isinstance(model_loader, RunaiModelStreamerLoader) | ||
|
||
def test_runai_model_loader_download_files(vllm_runner: VllmRunner): | ||
with vllm_runner( | ||
Check failure on line 27 in tests/runai_model_streamer/test_runai_model_streamer_loader.py GitHub Actions / mypy (3.9)
Check failure on line 27 in tests/runai_model_streamer/test_runai_model_streamer_loader.py GitHub Actions / mypy (3.10)
Check failure on line 27 in tests/runai_model_streamer/test_runai_model_streamer_loader.py GitHub Actions / mypy (3.11)
|
||
test_model, | ||
load_format=LoadFormat.RUNAI_STREAMER) as llm: | ||
deserialized_outputs = llm.generate( | ||
prompts, sampling_params) | ||
assert deserialized_outputs |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,38 @@ | ||
import glob | ||
import tempfile | ||
import torch | ||
|
||
import huggingface_hub.constants | ||
|
||
from vllm.model_executor.model_loader.weight_utils import ( | ||
download_weights_from_hf, runai_safetensors_weights_iterator, safetensors_weights_iterator) | ||
|
||
def test_runai_model_loader(): | ||
with tempfile.TemporaryDirectory() as tmpdir: | ||
huggingface_hub.constants.HF_HUB_OFFLINE = False | ||
download_weights_from_hf("openai-community/gpt2", | ||
allow_patterns=["*.safetensors"], | ||
cache_dir=tmpdir) | ||
safetensors = glob.glob(f"{tmpdir}/**/*.safetensors", recursive=True) | ||
assert len(safetensors) > 0 | ||
|
||
runai_model_streamer_tensors = {} | ||
hf_safetensors_tensors = {} | ||
|
||
for name, tensor in runai_safetensors_weights_iterator(safetensors): | ||
runai_model_streamer_tensors[name] = tensor | ||
|
||
for name, tensor in safetensors_weights_iterator(safetensors): | ||
hf_safetensors_tensors[name] = tensor | ||
|
||
assert len(runai_model_streamer_tensors) == len(hf_safetensors_tensors) | ||
|
||
for name, runai_tensor in runai_model_streamer_tensors.items(): | ||
assert runai_tensor.dtype == hf_safetensors_tensors[name].dtype | ||
assert runai_tensor.shape == hf_safetensors_tensors[name].shape | ||
assert torch.all(runai_tensor.eq(hf_safetensors_tensors[name])) | ||
|
||
|
||
|
||
if __name__ == "__main__": | ||
test_runai_model_loader() |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters