-
Notifications
You must be signed in to change notification settings - Fork 55
[WIP] Second pass at LaunchConfig #727
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
Auto-sync is disabled for draft pull requests in this repository. Workflows must be run manually. Contributors can view more details about this message here. |
cpcloud
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the PR!
This really needs some more explicit motivation in the PR description, as well as some real justification for all the duplicated tooling.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not entirely sure what the purpose of this file is beyond what's happening in the existing test_kernel_launch.py benchmarks.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think all of this can be done with the existing pytest-benchmark plugin.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think all or most of this functionality can be done with the existing pytest-benchmark plugin.
Really would like to avoid duplicating functionality, especially if it's AI generated duplication.
| - `bench-launch-overhead` | ||
| - `bench` | ||
| - `benchcmp` | ||
| - `bench-against` |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
While this script doesn't do a three way comparison, it also doesn't require writing any new code to run it.
Can we try to reuse bench-against instead of reinventing a lot of what that already does?
| def some_kernel_1(): | ||
| return | ||
|
|
||
| @cuda.jit("void(float32[:])") | ||
| def some_kernel_2(arr1): | ||
| return | ||
|
|
||
| @cuda.jit("void(float32[:],float32[:])") | ||
| def some_kernel_3(arr1, arr2): | ||
| return | ||
|
|
||
| @cuda.jit("void(float32[:],float32[:],float32[:])") | ||
| def some_kernel_4(arr1, arr2, arr3): | ||
| return | ||
|
|
||
| @cuda.jit("void(float32[:],float32[:],float32[:],float32[:])") | ||
| def some_kernel_5(arr1, arr2, arr3, arr4): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
These are nearly identical to the existing benchmarks. Let's avoid repeating existing benchmarks and tools that run them.
No description provided.