We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
arcee-ai/Virtuoso-Lite is a LlamaForCausalLM model, which I expected to be supported.
Compilation with LMI 0.30 fails:
2025-01-30 10:43:02.000490: 410 ERROR ||NEURON_CC_WRAPPER||: Failed compilation with ['neuronx-cc', 'compile', '--framework=XLA', '/tmp/no-user/neuroncc_compile_workdir/ff7ea03f-6b4c-4174-83e5-c9e847ff5520/model.MODULE_cc3503ed4b78fe12d0fa+54293761.hlo_module.pb', '--output', '/tmp/no-user/neuroncc_compile_workdir/ff7ea03f-6b4c-4174-83e5-c9e847ff5520/model.MODULE_cc3503ed4b78fe12d0fa+54293761.neff', '--target=trn1', '--logfile', '/tmp/compile.log', '--temp-dir=/tmp', '--model-type=transformer', '--auto-cast=none', '--execute-repetition=1', '--verbose=35']: 2025-01-30T10:43:02Z [F134] neuronx-cc terminated abnormally - Please open a support ticket at https://github.com/aws-neuron/aws-neuron-sdk/issues/new
How to reproduce:
DJL_IMAGE="763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.30.0-neuronx-sdk2.20.1" docker run -t --rm --network=host \ -v $PWD/model:/opt/ml/input/data/training \ $DEVICES \ $DJL_IMAGE \ partition --model-dir /opt/ml/input/data/training --skip-copy
serving.properties:
engine=Python option.dtype=bf16 option.entryPoint=djl_python.transformers_neuronx option.tensor_parallel_degree=24 option.n_positions=32768 option.max_rolling_batch_size=1 option.model_loading_timeout=3600 option.save_mp_checkpoint_path=/opt/ml/input/data/training/partition-test
The text was updated successfully, but these errors were encountered:
Hi @juliensimon,
Thanks for filing the issue. We will take a look and get back to you.
Sorry, something went wrong.
No branches or pull requests
arcee-ai/Virtuoso-Lite is a LlamaForCausalLM model, which I expected to be supported.
Compilation with LMI 0.30 fails:
How to reproduce:
serving.properties:
The text was updated successfully, but these errors were encountered: