From da4900e8359aba5923ac4f005608b292b77c45da Mon Sep 17 00:00:00 2001 From: klosax <131523366+klosax@users.noreply.github.com> Date: Mon, 31 Jul 2023 23:04:03 +0200 Subject: [PATCH] Update convert-llama-h5-to-gguf.py --- convert-llama-h5-to-gguf.py | 5 ----- 1 file changed, 5 deletions(-) diff --git a/convert-llama-h5-to-gguf.py b/convert-llama-h5-to-gguf.py index 412d334fb..7f1690595 100644 --- a/convert-llama-h5-to-gguf.py +++ b/convert-llama-h5-to-gguf.py @@ -104,11 +104,6 @@ if Path(dir_model + "/tokenizer.model").is_file(): print("Adding sentencepiece tokenizer vocab.") tokenizer = SentencePieceProcessor(dir_model + "/tokenizer.model") - # output vocab_size followed by all piece/score pairs - outbytes: bytes - outbytes = b"" - outbytes += struct.pack("I", tokenizer.vocab_size()) - for i in range(tokenizer.vocab_size()): text: bytes if tokenizer.is_unknown(i):