From 4f5b6224beb7b1a0291843cb118b148d5b6e7a6c Mon Sep 17 00:00:00 2001 From: klosax <131523366+klosax@users.noreply.github.com> Date: Mon, 31 Jul 2023 03:00:20 +0200 Subject: [PATCH] Update convert-gptneox-h5-to-gguf.py --- convert-gptneox-h5-to-gguf.py | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/convert-gptneox-h5-to-gguf.py b/convert-gptneox-h5-to-gguf.py index 12b6cc960..266199509 100644 --- a/convert-gptneox-h5-to-gguf.py +++ b/convert-gptneox-h5-to-gguf.py @@ -37,6 +37,12 @@ if len(sys.argv) > 2: sys.exit(1) fname_out = sys.argv[1] + "/ggml-model-" + ftype_str[ftype] + ".gguf" +with open(dir_model + "/config.json", "r", encoding="utf-8") as f: + hparams = json.load(f) + +if hparams["architectures"][0] != "GPTNeoXForCausalLM": + print("Model architecture not supported: " + hparams["architectures"][0] ) + sys.exit() model = AutoModelForCausalLM.from_pretrained(dir_model, low_cpu_mem_usage=True, trust_remote_code=True) list_vars = model.state_dict() @@ -51,9 +57,6 @@ for name in list_vars.keys(): gguf_writer = gguf.GGUFWriter.open(fname_out) -with open(dir_model + "/config.json", "r", encoding="utf-8") as f: - hparams = json.load(f) - # This must be changed when adding/deleting kv kv_count = 14