-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP: Add PARTED kernels #382
Open
MrBurmark
wants to merge
39
commits into
develop
Choose a base branch
from
feature/burmark1/parted
base: develop
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
MrBurmark
force-pushed
the
feature/burmark1/parted
branch
2 times, most recently
from
October 25, 2023 16:35
1536b7a
to
fd328c2
Compare
rhornung67
approved these changes
Oct 25, 2023
MrBurmark
force-pushed
the
feature/burmark1/parted
branch
from
November 21, 2023 17:44
1d5b8e4
to
9906d4c
Compare
This does the same thing as TRIAD but breaks it into multiple for loops over the data instead of a single for loop over the data.
Leave in comments of other dispatch options.
This makes each partition a multiple of the size of the prevoius partition
This tuning provides a best case scenario where the overhead of capturing the state and synchronizing per rep is removed.
The new gpu tuning is a AOS version using triad_holder. This is now in addition to the SOA tuning.
This copies the basic mempool from RAJA and adds a capability to synchronize as necessary to avoid host device race conditions when memory is needed on the host and but all the memory has been used on the device.
Default is on so the sizes of partitions are not always in non-decreasing order.
This uses a scan and binary search to schedule work to blocks instead of a 2d grid. Thus it avoids blocks with no work.
This is faster for cuda but slower for hip.
This has a minimal effect
with triad parted fused This has a large effect and makes a block size of 256 as good or better than 1024
always use binary search code
reorder TRIAD_PARTED gpu tuning declarations
These tuning use events to "fork-join" the streams as would be required in more realistic code. Though it would not always have to be done as frequently.
MrBurmark
force-pushed
the
feature/burmark1/parted
branch
from
January 31, 2024 19:55
97c41ec
to
f1d0120
Compare
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
WIP: Add PARTED kernels
These parted or partitioned kernels do the same computation over the same data as the Stream TRIAD kernel, but instead of a single loop over all of the data they use multiple kernels over parts of the data. So the same work is ultimately done to the same data just broken up into multiple partitions.
The idea is to look at ways of improving performance when running the same kernel over different data like when running over subdomains or tiles in a block structured code or with AMR.
Ideas:
add kernels listed in Add kernels broken up over multiple loops #381
add a folder to keep this performance study in a single place
split up data unevenly (currently each part gets size/num_parts data)
This PR is a feature
It does the following: