diff options
-rw-r--r-- | bitsandbytes/autograd/_functions.py | 2 |
1 files changed, 2 insertions, 0 deletions
diff --git a/bitsandbytes/autograd/_functions.py b/bitsandbytes/autograd/_functions.py index 9928fbd..407f14b 100644 --- a/bitsandbytes/autograd/_functions.py +++ b/bitsandbytes/autograd/_functions.py @@ -370,6 +370,8 @@ class MatMul8bitLt(torch.autograd.Function): if state.threshold > 0.0 and subA is not None: grad_B[:, idx] += torch.matmul(grad_output.t(), subA) + raise NotImplementedError("!!") + if req_gradA: if state.CBt is not None: C32grad, Sgrad = F.transform(Cgrad, "col32") |