mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
Better error message on non rocm.
This commit is contained in:
parent
a76821e0b2
commit
326f8e30ac
@ -354,6 +354,8 @@ def get_linear(weight, bias, quantize):
|
|||||||
"AWQ GEMM kernel can't be used on ROCm systems, please use `--quantize gptq` instead "
|
"AWQ GEMM kernel can't be used on ROCm systems, please use `--quantize gptq` instead "
|
||||||
"to use Exllama/GPTQ kernels for AWQ inference."
|
"to use Exllama/GPTQ kernels for AWQ inference."
|
||||||
)
|
)
|
||||||
|
if not HAS_AWQ:
|
||||||
|
raise NotImplementedError("You do not seem to have awq installed, either install it (cd server && make install-awq), or try using GPTQ `---quantize gptq` a conversion AWQ->GPTQ will happen on the fly")
|
||||||
linear = WQLinear(
|
linear = WQLinear(
|
||||||
w_bit=bits,
|
w_bit=bits,
|
||||||
group_size=groupsize,
|
group_size=groupsize,
|
||||||
|
Loading…
Reference in New Issue
Block a user