You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have followed the readme file to config all of the setup steps, including downloading the dataset. When I directly run the inference command, the output the the model is random characters.
Some setup steps:
(1)Environment installation is same as the requirements, including the specific version of transformer.
(2)The original LLaMA weights are downloaded from HuggingFace website and using the official conversion command. Then applying the shikras/shikra-7b-delta-v1 to the original weights.
(3)Download the dataset images used in the repo and change the dataset root. For inference stage, I use the shikra_eval_multi_pope script, the default configuration file is 'DEFAULT_TEST_POPE_VARIANT', the dataset used is COCO val2014 dataset.
The prediction is either empty or garbled output in the output_dir/multitest_xxxx_extra_prediction.jsonl.
The metric computation shows all of the results are false, like:
I have followed the
readme
file to config all of the setup steps, including downloading the dataset. When I directly run the inference command, the output the the model is random characters.Some setup steps:
(1)Environment installation is same as the requirements, including the specific version of transformer.
(2)The original LLaMA weights are downloaded from HuggingFace website and using the official conversion command. Then applying the
shikras/shikra-7b-delta-v1
to the original weights.(3)Download the dataset images used in the repo and change the dataset root. For inference stage, I use the
shikra_eval_multi_pope
script, the default configuration file is 'DEFAULT_TEST_POPE_VARIANT', the dataset used is COCO val2014 dataset.The command I use for the inference is:
using a single NVIDIA A100 GPU.
But the output for
COCO_POPE_RANDOM_q_a
,COCO_POPE_POPULAR_q_a
andCOCO_POPE_ADVERSARIAL_q_a
, all of the output of the model is like:or
The prediction is either empty or garbled output in the
output_dir/multitest_xxxx_extra_prediction.jsonl
.The metric computation shows all of the results are false, like:
I check all of the configurations and didn't find some errors. So could you please give me some suggestions? Thanks!
The text was updated successfully, but these errors were encountered: