You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have implemented a simple example where I want to do hp tuning using optuna and each trial is spawn over two gpus (as my original data is huge). I am not looking for parallelizing trial runs themselves rather doing data parellism across 2 gpus within each trial. I expected that the trial.report() would work from custom PruningCallback I implemented, hwoever it's not pruning. I am trying the replicate the result (same trials should get pruned) when running on a single GPU with everythign else the same.
Environment
Optuna version: 2.10.1
Python version: 3.10.15
OS: Linux
torch: 2.2.2+cu121
Error messages, stack traces, or logs
When I run mlp_ddp.py script below none of the trials get pruned as a result of PruningCallback not working properly with optuna
Steps to reproduce
The mlp_ddp.py code contains implementation with ddp within each trial (expect that same trials as "mlp.py" script to be pruned, but don't see any trials getting pruned)
mlp.py code below contains implementation for a single gpu (reference solution for which trials should be pruned)
Expected behavior
I have implemented a simple example where I want to do hp tuning using optuna and each trial is spawn over two gpus (as my original data is huge). I am not looking for parallelizing trial runs themselves rather doing data parellism across 2 gpus within each trial. I expected that the trial.report() would work from custom PruningCallback I implemented, hwoever it's not pruning. I am trying the replicate the result (same trials should get pruned) when running on a single GPU with everythign else the same.
Environment
Error messages, stack traces, or logs
When I run mlp_ddp.py script below none of the trials get pruned as a result of PruningCallback not working properly with optuna
Steps to reproduce
Reproducible examples (optional)
Additional context (optional)
The text was updated successfully, but these errors were encountered: