From 6e43ef51bac007f6bb85904eaf8e7acd2b162f35 Mon Sep 17 00:00:00 2001 From: OlivierDehaene <23298448+OlivierDehaene@users.noreply.github.com> Date: Thu, 26 Jan 2023 15:33:14 +0100 Subject: [PATCH] feat(bloom): use torch.nn.Linear and torch.nn.GELU --- server/text_generation/models/bloom.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/server/text_generation/models/bloom.py b/server/text_generation/models/bloom.py index 2218b91b..463c0406 100644 --- a/server/text_generation/models/bloom.py +++ b/server/text_generation/models/bloom.py @@ -136,7 +136,6 @@ class BLOOMSharded(BLOOM): start = rank * block_size stop = (rank + 1) * block_size tensor = slice_[start:stop] - tensor = tensor.transpose(1, 0) else: size = slice_.get_shape()[0] block_size = size // world_size @@ -150,7 +149,6 @@ class BLOOMSharded(BLOOM): start = rank * block_size stop = (rank + 1) * block_size tensor = slice_[:, start:stop] - tensor = tensor.transpose(1, 0) else: tensor = slice_[:] # XXX: Hack for Rowlinear to add the bias only once. @@ -186,7 +184,7 @@ class BLOOMSharded(BLOOM): and param_name == "weight" ): tensor = Int8Params( - tensor.transpose(1, 0), + tensor, has_fp16_weights=False, requires_grad=False, ).to(device)