Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU optimization #1

Open
colindaven opened this issue Jun 21, 2021 · 2 comments
Open

GPU optimization #1

colindaven opened this issue Jun 21, 2021 · 2 comments

Comments

@colindaven
Copy link
Owner

From @maxdeest

runners * chunks_per_runner * chunk_size ~= 100000 * [max GPU memory in GB] * 2

For Ampere A100 GPU, max out the 40 GB GPU mem
--chunk_size 3000 --gpu_runners_per_device 8 --chunks_per_runner 512

@BeneKenobi
Copy link

Why do you use less gpu_runners_per_device in

gpu_params='--compress_fastq --num_callers 4 --gpu_runners_per_device 4 --chunks_per_runner 512 --chunk_size 3000'
? Where there problems with more runners?

@colindaven
Copy link
Owner Author

colindaven commented Mar 10, 2023

I don't think there was a problem with more runners, it was maybe just set for a different GPU. You need to play with the settings to reach the optimal speed for your GPU, and not everyone has a A100 with 40 GB GPU RAM.

Just monitor the GPU device usage and RAM using nvidia-smi.

Remember IO has a big impact, so storing fast5 files on SSD or other fast storage can decrease runtime a lot (36h to 24h for a human genome).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants