-
Notifications
You must be signed in to change notification settings - Fork 2.9k
[feat] fix some numel and refactor torch binding #4867
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This PR do three things
from: m.def("all_reduce(int fa, Tensor inp, Tensor! out) -> ()");
m.impl("all_reduce", torch::kCUDA, &all_reduce); to: m.def("all_reduce", all_reduce); simplify the kernel define logic
|
@@ -4,9 +4,9 @@ | |||
[submodule "sgl-kernel/3rdparty/cccl"] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We no longer need to update git modules because we have switched to CMake.
GIT_REPOSITORY https://github.com/flashinfer-ai/flashinfer | ||
GIT_TAG 79fd1ae90d9b8098ca70dec6071da96f3f6da7b9 | ||
GIT_REPOSITORY https://github.com/sgl-project/flashinfer | ||
GIT_TAG 2b9f16eb79bd344e31725e8d7a92fe7fe980ffdf |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please use branch instead.
@@ -22,181 +22,67 @@ TORCH_LIBRARY_EXPAND(sgl_kernel, m) { | |||
/* | |||
* From csrc/allreduce | |||
*/ | |||
m.def( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
revert this first in #4871 to unblock the new release
Motivation
fix some numel and refactor torch binding
Modifications
Checklist