You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The commands look correct. I'm just not sure if the difference between the model name in the export and deployment is accidental.
I assume the command to export model was:
I tested manually that this model work fine in ovms.
The error message from your log suggest that the model in $MODEL_DIR/HuggingFaceTB/SmolLM2-135M-Instruct is invalid. Could you send the output of ls -l $MODEL_DIR/HuggingFaceTB/SmolLM2-135M-Instruct
Describe the bug
A clear and concise description of what the bug is.
I have prepared a
text-generation
model using the filedemos/common/export_models/export_model.py
. The config file is:When I run the inference server using the docker container:
sudo docker run \ --rm -d \ -p 8085:8085 \ -v $MODEL_DIR:/workspace:ro \ openvino/model_server:2024.5 \ --rest_port 8085 \ --rest_bind_address 0.0.0.0 \ --config_path /workspace/config.json
The server starts but i also get an error:
To Reproduce
Steps to reproduce the behavior:
Run the command:
Run the docker image:
sudo docker run \ --rm -d \ -p 8085:8085 \ -v $MODEL_DIR:/workspace:ro \ openvino/model_server:2024.5 \ --rest_port 8085 \ --rest_bind_address 0.0.0.0 \ --config_path /workspace/config.json --log_level DEBUG
Expected behavior
Expected behaviour is for the server to start and to be able to respond to the requests.
Configuration
I need help with identifying any mistakes that I am doing during preparation and running the docker container.
The text was updated successfully, but these errors were encountered: