mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 20:34:54 +00:00
fix
This commit is contained in:
parent
f2fecdceca
commit
61b49859da
@ -18,7 +18,6 @@ vllm-rocm:
|
||||
|
||||
build-vllm-rocm: vllm-rocm
|
||||
cd vllm && git fetch && git checkout ca6913b3c2ffacdcb7d15e914dc34adbc6c89479
|
||||
cd vllm && patch /opt/rocm/include/hip/amd_detail/amd_hip_bf16.h ./rocm_patch/rocm_bf16.patch
|
||||
cd vllm && PYTORCH_ROCM_ARCH="gfx90a;gfx942" python setup.py install
|
||||
|
||||
install-vllm-rocm: build-vllm-rocm
|
||||
|
@ -244,7 +244,7 @@ class LlamaMLP(nn.Module):
|
||||
)
|
||||
|
||||
def forward(self, hidden_states):
|
||||
if False and IS_ROCM_SYSTEM and self.hidden_act == "silu" and hidden_states.shape[0] == 1:
|
||||
if IS_ROCM_SYSTEM and self.hidden_act == "silu" and hidden_states.shape[0] == 1:
|
||||
out = torch.empty(
|
||||
hidden_states.shape[0],
|
||||
self.intermediate_size,
|
||||
|
@ -366,7 +366,7 @@ class FastLinearROCm(nn.Module):
|
||||
weight = self.weight
|
||||
bias = self.bias
|
||||
|
||||
if False and IS_ROCM_SYSTEM and inp.numel() // inp.size(-1) == 1:
|
||||
if IS_ROCM_SYSTEM and inp.numel() // inp.size(-1) == 1:
|
||||
batched = False
|
||||
|
||||
if inp.dim() == 3:
|
||||
|
Loading…
Reference in New Issue
Block a user