mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-22 15:32:08 +00:00
fix: allocate tmp based on sgmv kernel if available (#2345)
* fix: allocate tmp based on sgmv kernel if available * fix: re add copy build artifacts step for punica kernels
This commit is contained in:
parent
8e6bfa2fc5
commit
3079865b60
@ -151,13 +151,17 @@ def get_tmp_expand_size(size: int) -> int:
|
||||
def get_tmp_tensors(
|
||||
nsegments: int, lora_rank: int, device: torch.device
|
||||
) -> Tuple[torch.Tensor, torch.Tensor]:
|
||||
if use_cutlass_shrink(lora_rank) and has_sgmv():
|
||||
use_cutlass = use_cutlass_shrink(lora_rank) and has_sgmv()
|
||||
has_sgmv_available = has_sgmv()
|
||||
|
||||
if use_cutlass:
|
||||
tmp = get_tmp_tensor_for_size(nsegments, device)
|
||||
return tmp, tmp
|
||||
elif has_sgmv_available:
|
||||
return get_tmp_tensor(device), get_tmp_tensor_for_size(nsegments, device)
|
||||
else:
|
||||
tmp_shrink = get_tmp_tensor(device)
|
||||
tmp_expand = get_tmp_tensor_for_size_no_kernels(nsegments, device)
|
||||
return tmp_shrink, tmp_expand
|
||||
tmp = get_tmp_tensor_for_size(nsegments, device)
|
||||
return tmp, tmp
|
||||
|
||||
|
||||
def lora_a_sgmv_cutlass(
|
||||
|
Loading…
Reference in New Issue
Block a user