mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-09 19:34:53 +00:00
fmt
This commit is contained in:
parent
1111125092
commit
e4ad3066bc
@ -39,7 +39,9 @@ class FlashLlama(FlashCausalLM):
|
|||||||
raise NotImplementedError("FlashLlama does not support quantization")
|
raise NotImplementedError("FlashLlama does not support quantization")
|
||||||
|
|
||||||
tokenizer = LlamaTokenizer.from_pretrained(
|
tokenizer = LlamaTokenizer.from_pretrained(
|
||||||
model_id, revision=revision, padding_side="left",
|
model_id,
|
||||||
|
revision=revision,
|
||||||
|
padding_side="left",
|
||||||
)
|
)
|
||||||
|
|
||||||
config = AutoConfig.from_pretrained(
|
config = AutoConfig.from_pretrained(
|
||||||
@ -155,7 +157,9 @@ class FlashLlamaSharded(FlashLlama):
|
|||||||
raise NotImplementedError("FlashLlama does not support quantization")
|
raise NotImplementedError("FlashLlama does not support quantization")
|
||||||
|
|
||||||
tokenizer = LlamaTokenizer.from_pretrained(
|
tokenizer = LlamaTokenizer.from_pretrained(
|
||||||
model_id, revision=revision, padding_side="left",
|
model_id,
|
||||||
|
revision=revision,
|
||||||
|
padding_side="left",
|
||||||
)
|
)
|
||||||
|
|
||||||
config = AutoConfig.from_pretrained(
|
config = AutoConfig.from_pretrained(
|
||||||
|
Loading…
Reference in New Issue
Block a user