Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance for summarization task on BART is low after latest Transformer 4.40 upgrade #1144

Open
2 of 4 tasks
astachowiczhabana opened this issue Jul 19, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@astachowiczhabana
Copy link
Contributor

System Info

Bad
Optimum Habana latest main: c495f479d9abf04fb7adb6f0a5607d7963186649
Synapse docker image: v1.16

Good:
Optimum Habana one commit before Transformer 4.40 upgrade: 569580ff9bf44083514533ad28e336043891947b
Synapse docker image: v1.16

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

cd /root/optimum-habana/examples/summarization
pip install -r requirements.txt
PT_HPU_ENABLE_REFINE_DYNAMIC_SHAPES=1 python run_summarization.py --model_name_or_path facebook/bart-large-cnn --do_predict --predict_with_generate --dataset_name cnn_dailymail --dataset_config "3.0.0" --output_dir ./tst-summarization --overwrite_output_dir --per_device_eval_batch_size 2 --use_habana --use_lazy_mode --use_hpu_graphs_for_inference --gaudi_config_name Habana/t5 --ignore_pad_token_for_loss False --pad_to_max_length --num_beams 1 --generation_num_beams 1 --bf16 --ignore_eos False

Expected behavior

The quickest way to check if something is wrong is observe performance.

Before Transformer 4.40 upgrade the speed is ~3.9 it/s
After Transformer 4.40 upgrade the speed is ~1.7 it/s

@astachowiczhabana astachowiczhabana added the bug Something isn't working label Jul 19, 2024
@regisss
Copy link
Collaborator

regisss commented Oct 22, 2024

@astachowiczhabana Are we still having this regression?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants