You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We shouldn't have to search for the kernel in the module every run; let's just find it once and keep it in the CUDA-specific member of the execution context, explicitly. This would probably let us avoid having to keep the mangled name there.
The text was updated successfully, but these errors were encountered:
…rnel wrappers
* Not search for the mangled kernel by name with every run
* Execution context now explicitly holds a `cuda::kernel_t`
* Execution context no longer holds the mangled kernel name
* Now printing the kernel's amount of static shared memory used, after the build
We shouldn't have to search for the kernel in the module every run; let's just find it once and keep it in the CUDA-specific member of the execution context, explicitly. This would probably let us avoid having to keep the mangled name there.
The text was updated successfully, but these errors were encountered: