mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-27 20:04:35 +00:00
Update convert-gptneox-h5-to-gguf.py
This commit is contained in:
parent
2a0914673c
commit
4f5b6224be
@ -37,6 +37,12 @@ if len(sys.argv) > 2:
|
|||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
fname_out = sys.argv[1] + "/ggml-model-" + ftype_str[ftype] + ".gguf"
|
fname_out = sys.argv[1] + "/ggml-model-" + ftype_str[ftype] + ".gguf"
|
||||||
|
|
||||||
|
with open(dir_model + "/config.json", "r", encoding="utf-8") as f:
|
||||||
|
hparams = json.load(f)
|
||||||
|
|
||||||
|
if hparams["architectures"][0] != "GPTNeoXForCausalLM":
|
||||||
|
print("Model architecture not supported: " + hparams["architectures"][0] )
|
||||||
|
sys.exit()
|
||||||
|
|
||||||
model = AutoModelForCausalLM.from_pretrained(dir_model, low_cpu_mem_usage=True, trust_remote_code=True)
|
model = AutoModelForCausalLM.from_pretrained(dir_model, low_cpu_mem_usage=True, trust_remote_code=True)
|
||||||
list_vars = model.state_dict()
|
list_vars = model.state_dict()
|
||||||
@ -51,9 +57,6 @@ for name in list_vars.keys():
|
|||||||
|
|
||||||
gguf_writer = gguf.GGUFWriter.open(fname_out)
|
gguf_writer = gguf.GGUFWriter.open(fname_out)
|
||||||
|
|
||||||
with open(dir_model + "/config.json", "r", encoding="utf-8") as f:
|
|
||||||
hparams = json.load(f)
|
|
||||||
|
|
||||||
# This must be changed when adding/deleting kv
|
# This must be changed when adding/deleting kv
|
||||||
kv_count = 14
|
kv_count = 14
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user