mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 20:34:54 +00:00
Ignore missing generation config.
This commit is contained in:
parent
1f1885d911
commit
ffea15d6b6
@ -52,12 +52,15 @@ class FlashLlama(FlashCausalLM):
|
|||||||
truncation_side="left",
|
truncation_side="left",
|
||||||
trust_remote_code=trust_remote_code,
|
trust_remote_code=trust_remote_code,
|
||||||
)
|
)
|
||||||
generation_config = GenerationConfig.from_pretrained(
|
try:
|
||||||
model_id, revision=revision, trust_remote_code=trust_remote_code
|
generation_config = GenerationConfig.from_pretrained(
|
||||||
)
|
model_id, revision=revision, trust_remote_code=trust_remote_code
|
||||||
if isinstance(generation_config.eos_token_id, (list, set)):
|
)
|
||||||
# TODO Huge hack
|
if isinstance(generation_config.eos_token_id, (list, set)):
|
||||||
tokenizer._eos_token_ids = set(generation_config.eos_token_id)
|
# TODO Huge hack
|
||||||
|
tokenizer._eos_token_ids = set(generation_config.eos_token_id)
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
config = AutoConfig.from_pretrained(
|
config = AutoConfig.from_pretrained(
|
||||||
model_id, revision=revision, trust_remote_code=trust_remote_code
|
model_id, revision=revision, trust_remote_code=trust_remote_code
|
||||||
|
Loading…
Reference in New Issue
Block a user