Skip to content

Commit

Permalink
Refactor fbgemm / llama csrc code base
Browse files Browse the repository at this point in the history
Summary: Dedup and reuse the ops moved in fbgemm gpu.

Reviewed By: sryap, xintwfb, jiawenliu64

Differential Revision: D56685973

fbshipit-source-id: e3e6e27e9f3f380460f4fbf01eb3075109e37d2f
  • Loading branch information
jianyuh authored and facebook-github-bot committed Apr 30, 2024
1 parent ca4e84b commit 0da2f0c
Showing 1 changed file with 1 addition and 1 deletion.
Original file line number Diff line number Diff line change
Expand Up @@ -189,7 +189,7 @@ def test_tensor_with_nan(self, G_B: int, D: int, HD_L: int) -> None:
w = wq.bfloat16() * w_scale

zq_ref = (x @ w.T).to(torch.bfloat16)
torch.testing.assert_close(zq[:B, :], zq_ref, atol=1.0e-3, rtol=1.0e-3)
torch.testing.assert_close(zq[:B, :], zq_ref, atol=2.0e-3, rtol=2.0e-3)

@settings(deadline=None)
@given(
Expand Down

0 comments on commit 0da2f0c

Please sign in to comment.