From 67a1283501fa24d346f8e8efb4fc888a9ed8d193 Mon Sep 17 00:00:00 2001 From: Robin Schmidt Date: Mon, 15 Nov 2021 17:27:02 +0100 Subject: [FIX] passing of sparse in StableEmbedding --- bitsandbytes/nn/modules.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'bitsandbytes/nn/modules.py') diff --git a/bitsandbytes/nn/modules.py b/bitsandbytes/nn/modules.py index bf0945c..ce2f3a4 100644 --- a/bitsandbytes/nn/modules.py +++ b/bitsandbytes/nn/modules.py @@ -15,8 +15,8 @@ from bitsandbytes.optim import GlobalOptimManager class StableEmbedding(torch.nn.Embedding): def __init__(self, num_embeddings: int, embedding_dim: int, padding_idx: Optional[int] = None, max_norm: Optional[float] = None, norm_type: float = 2., scale_grad_by_freq: bool = False, - sparse: bool = True, _weight: Optional[Tensor] = None) -> None: - super(StableEmbedding, self).__init__(num_embeddings, embedding_dim, padding_idx, max_norm, norm_type, scale_grad_by_freq, False, _weight) + sparse: bool = False, _weight: Optional[Tensor] = None) -> None: + super(StableEmbedding, self).__init__(num_embeddings, embedding_dim, padding_idx, max_norm, norm_type, scale_grad_by_freq, sparse, _weight) self.norm = torch.nn.LayerNorm(embedding_dim) GlobalOptimManager.get_instance().register_parameters(self.weight) GlobalOptimManager.get_instance().override_config(self.weight, 'optim_bits', 32) -- cgit v1.2.3