Skip to content

Commit

Permalink
Revert D51999938: Multisect successfully blamed "D51999938: Add early…
Browse files Browse the repository at this point in the history
… exit in sparse_async_cumsum ops" for otest failure (#2208)

Summary:
Pull Request resolved: #2208

This diff is reverting D51999938
D51999938: Add early exit in sparse_async_cumsum ops by meremeev has been identified to be causing the following test failure:

Tests affected:
- [deeplearning/fbgemm/fbgemm_gpu:sparse_ops_test - test_schema__test_asynchronous_complete_cumsum_2d (deeplearning.fbgemm.fbgemm_gpu.test.sparse_ops_test.SparseOpsTest)](https://www.internalfb.com/intern/test/562950068047718/)

Here's the Multisect link:
https://www.internalfb.com/multisect/3757962
Here are the tasks that are relevant to this breakage:

We're generating a revert to back out the changes in this diff, please note the backout may land if someone accepts it.

If you believe this diff has been generated in error you may Commandeer and Abandon it.

Reviewed By: jasonjk-park

Differential Revision: D52099677

fbshipit-source-id: 4e36745864148e5bb337465b0a9afcfe80846389
  • Loading branch information
generatedunixname89002005232357 authored and facebook-github-bot committed Dec 13, 2023
1 parent 157f935 commit af058b0
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 12 deletions.
9 changes: 0 additions & 9 deletions fbgemm_gpu/src/sparse_ops/sparse_async_cumsum.cu
Original file line number Diff line number Diff line change
Expand Up @@ -13,9 +13,6 @@ using Tensor = at::Tensor;
namespace fbgemm_gpu {

DLL_PUBLIC Tensor asynchronous_inclusive_cumsum_gpu(const Tensor& t_in) {
if (t_in.numel() == 0) {
return at::zeros_like(t_in);
}
TENSOR_ON_CUDA_GPU(t_in);

at::cuda::OptionalCUDAGuard device_guard;
Expand Down Expand Up @@ -57,9 +54,6 @@ DLL_PUBLIC Tensor asynchronous_inclusive_cumsum_gpu(const Tensor& t_in) {
}

DLL_PUBLIC Tensor asynchronous_exclusive_cumsum_gpu(const Tensor& t_in) {
if (t_in.numel() == 0) {
return at::zeros_like(t_in);
}
TENSOR_ON_CUDA_GPU(t_in);

at::cuda::OptionalCUDAGuard device_guard;
Expand Down Expand Up @@ -101,9 +95,6 @@ DLL_PUBLIC Tensor asynchronous_exclusive_cumsum_gpu(const Tensor& t_in) {
}

DLL_PUBLIC Tensor asynchronous_complete_cumsum_gpu(const Tensor& t_in) {
if (t_in.numel() == 0) {
return at::zeros({t_in.numel() + 1}, t_in.options());
}
TENSOR_ON_CUDA_GPU(t_in);

at::cuda::OptionalCUDAGuard device_guard;
Expand Down
6 changes: 3 additions & 3 deletions fbgemm_gpu/test/sparse_ops_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -591,7 +591,7 @@ def test_block_bucketize_sparse_features_long_indices(
torch.testing.assert_close(new_indices_gpu.cpu(), new_indices_cpu)

@given(
n=st.integers(min_value=0, max_value=100),
n=st.integers(min_value=1, max_value=100),
long_index=st.booleans(),
)
@settings(verbosity=Verbosity.verbose, max_examples=20, deadline=None)
Expand Down Expand Up @@ -655,8 +655,8 @@ def test_cumsum(self, n: int, long_index: bool) -> None:
)

@given(
n=st.integers(min_value=0, max_value=600),
b=st.integers(min_value=0, max_value=10),
n=st.integers(min_value=1, max_value=600),
b=st.integers(min_value=1, max_value=10),
long_index=st.booleans(),
)
@settings(verbosity=Verbosity.verbose, max_examples=20, deadline=None)
Expand Down

0 comments on commit af058b0

Please sign in to comment.