Local gptq support.

This commit is contained in:
Nicolas Patry 2023-07-31 09:51:58 +02:00
parent 3ef5ffbc64
commit f29e3d7d34

View File

@ -1,3 +1,4 @@
import os
from pathlib import Path
from typing import List, Dict, Optional, Tuple
from safetensors import safe_open, SafetensorError
@ -221,8 +222,12 @@ class Weights:
return bits, groupsize
def _set_gptq_params(self, model_id):
filename = "quantize_config.json"
try:
filename = hf_hub_download(model_id, filename="quantize_config.json")
if not os.path.exists(os.path.join(model_id, filename)):
filename = os.path.join(model_id, filename)
else:
filename = hf_hub_download(model_id, filename=filename)
with open(filename, "r") as f:
data = json.load(f)
self.gptq_bits = data["bits"]