2024-04-29 13:58:41 +00:00
# This script downloads the tokenizer models of the specified models from Huggingface and
# generates the get_vocab_base_pre() function for convert-hf-to-gguf.py
#
# This is necessary in order to analyze the type of pre-tokenizer used by the model and
# provide the necessary information to llama.cpp via the GGUF header in order to implement
# the same pre-tokenizer.
#
# ref: https://github.com/ggerganov/llama.cpp/pull/6920
#
# Instructions:
#
# - Add a new model to the "models" list
# - Run the script with your huggingface token:
#
# python3 convert-hf-to-gguf-update.py <huggingface_token>
#
# - Copy-paste the generated get_vocab_base_pre() function into convert-hf-to-gguf.py
# - Update llama.cpp with the new pre-tokenizer if necessary
#
# TODO: generate tokenizer tests for llama.cpp
# TODO: automate the update of convert-hf-to-gguf.py
#
2024-05-03 19:36:41 +00:00
import logging
2024-04-29 13:58:41 +00:00
import os
import requests
import sys
import json
from hashlib import sha256
from enum import IntEnum , auto
2024-05-03 19:36:41 +00:00
from transformers import AutoTokenizer
logger = logging . getLogger ( " convert-hf-to-gguf-update " )
2024-04-29 13:58:41 +00:00
class TOKENIZER_TYPE ( IntEnum ) :
SPM = auto ( )
BPE = auto ( )
WPM = auto ( )
2024-05-03 19:36:41 +00:00
2024-04-29 13:58:41 +00:00
# TODO: this string has to exercise as much pre-tokenizer functionality as possible
# will be updated with time - contributions welcome
chktxt = ' \n \n \n \n \n \n \t \t \t \t \n \n \n \n \n 🚀 (normal) 😶🌫️ (multiple emojis concatenated) ✅ 🦙🦙 3 33 333 3333 33333 333333 3333333 33333333 3.3 3..3 3...3 កាន់តែពិសេសអាច😁 ?我想在apple工作1314151天~ ------======= нещо на Български \' \' \' \' \' \' ``````` \" \" \" \" ......!!!!!!?????? I \' ve been \' told he \' s there, \' RE you sure? \' M not sure I \' ll make it, \' D you like some tea? We \' Ve a \' lL '
if len ( sys . argv ) == 2 :
token = sys . argv [ 1 ]
else :
2024-05-03 19:36:41 +00:00
logger . info ( " Usage: python convert-hf-to-gguf-update.py <huggingface_token> " )
2024-04-29 13:58:41 +00:00
sys . exit ( 1 )
# TODO: add models here, base models preferred
models = [
2024-05-03 19:36:41 +00:00
{ " name " : " llama-spm " , " tokt " : TOKENIZER_TYPE . SPM , " repo " : " https://huggingface.co/meta-llama/Llama-2-7b-hf " , } ,
{ " name " : " llama-bpe " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/meta-llama/Meta-Llama-3-8B " , } ,
{ " name " : " phi-3 " , " tokt " : TOKENIZER_TYPE . SPM , " repo " : " https://huggingface.co/microsoft/Phi-3-mini-4k-instruct " , } ,
{ " name " : " deepseek-llm " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/deepseek-ai/deepseek-llm-7b-base " , } ,
{ " name " : " deepseek-coder " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-base " , } ,
{ " name " : " falcon " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/tiiuae/falcon-7b " , } ,
{ " name " : " bert-bge " , " tokt " : TOKENIZER_TYPE . WPM , " repo " : " https://huggingface.co/BAAI/bge-small-en-v1.5 " , } ,
{ " name " : " mpt " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/mosaicml/mpt-7b " , } ,
{ " name " : " starcoder " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/bigcode/starcoder2-3b " , } ,
{ " name " : " gpt-2 " , " tokt " : TOKENIZER_TYPE . BPE , " repo " : " https://huggingface.co/openai-community/gpt2 " , } ,
]
2024-04-29 13:58:41 +00:00
# make directory "models/tokenizers" if it doesn't exist
if not os . path . exists ( " models/tokenizers " ) :
os . makedirs ( " models/tokenizers " )
2024-05-03 19:36:41 +00:00
2024-04-29 13:58:41 +00:00
def download_file_with_auth ( url , token , save_path ) :
headers = { " Authorization " : f " Bearer { token } " }
response = requests . get ( url , headers = headers )
if response . status_code == 200 :
with open ( save_path , ' wb ' ) as f :
f . write ( response . content )
2024-05-03 19:36:41 +00:00
logger . info ( f " File { save_path } downloaded successfully " )
2024-04-29 13:58:41 +00:00
else :
2024-05-03 19:36:41 +00:00
logger . info ( f " Failed to download file. Status code: { response . status_code } " )
2024-04-29 13:58:41 +00:00
# download the tokenizer models
for model in models :
name = model [ " name " ]
repo = model [ " repo " ]
tokt = model [ " tokt " ]
if not os . path . exists ( f " models/tokenizers/ { name } " ) :
os . makedirs ( f " models/tokenizers/ { name } " )
else :
2024-05-03 19:36:41 +00:00
logger . info ( f " Directory models/tokenizers/ { name } already exists - skipping " )
2024-04-29 13:58:41 +00:00
continue
2024-05-03 19:36:41 +00:00
logger . info ( f " Downloading { name } to models/tokenizers/ { name } " )
2024-04-29 13:58:41 +00:00
url = f " { repo } /raw/main/config.json "
save_path = f " models/tokenizers/ { name } /config.json "
download_file_with_auth ( url , token , save_path )
url = f " { repo } /raw/main/tokenizer.json "
save_path = f " models/tokenizers/ { name } /tokenizer.json "
download_file_with_auth ( url , token , save_path )
if tokt == TOKENIZER_TYPE . SPM :
url = f " { repo } /resolve/main/tokenizer.model "
save_path = f " models/tokenizers/ { name } /tokenizer.model "
download_file_with_auth ( url , token , save_path )
url = f " { repo } /raw/main/tokenizer_config.json "
save_path = f " models/tokenizers/ { name } /tokenizer_config.json "
download_file_with_auth ( url , token , save_path )
# generate the source code for the convert-hf-to-gguf.py:get_vocab_base_pre() function:
# TODO: auto-update convert-hf-to-gguf.py with the generated function
src_ifs = " "
for model in models :
name = model [ " name " ]
tokt = model [ " tokt " ]
if tokt == TOKENIZER_TYPE . SPM :
continue
# create the tokenizer
tokenizer = AutoTokenizer . from_pretrained ( f " models/tokenizers/ { name } " )
chktok = tokenizer . encode ( chktxt )
chkhsh = sha256 ( str ( chktok ) . encode ( ) ) . hexdigest ( )
2024-05-03 19:36:41 +00:00
logger . info ( f " model: { name } " )
logger . info ( f " tokt: { tokt } " )
logger . info ( f " repo: { model [ ' repo ' ] } " )
logger . info ( f " chktok: { chktok } " )
logger . info ( f " chkhsh: { chkhsh } " )
2024-04-29 13:58:41 +00:00
# print the "pre_tokenizer" content from the tokenizer.json
2024-04-30 08:05:25 +00:00
with open ( f " models/tokenizers/ { name } /tokenizer.json " , " r " , encoding = " utf-8 " ) as f :
2024-04-29 13:58:41 +00:00
cfg = json . load ( f )
pre_tokenizer = cfg [ " pre_tokenizer " ]
2024-05-03 19:36:41 +00:00
logger . info ( " pre_tokenizer: " + json . dumps ( pre_tokenizer , indent = 4 ) )
2024-04-29 13:58:41 +00:00
2024-05-03 19:36:41 +00:00
logger . info ( " " )
2024-04-29 13:58:41 +00:00
src_ifs + = f " if chkhsh == \" { chkhsh } \" : \n "
src_ifs + = f " # ref: { model [ ' repo ' ] } \n "
src_ifs + = f " res = \" { name } \" \n "
2024-05-03 19:36:41 +00:00
src_func = f """
def get_vocab_base_pre ( self , tokenizer ) - > str :
# encoding this string and hashing the resulting tokens would (hopefully) give us a unique identifier that
# is specific for the BPE pre-tokenizer used by the model
# we will use this unique identifier to write a "tokenizer.ggml.pre" entry in the GGUF file which we can
# use in llama.cpp to implement the same pre-tokenizer
chktxt = { repr ( chktxt ) }
chktok = tokenizer . encode ( chktxt )
chkhsh = sha256 ( str ( chktok ) . encode ( ) ) . hexdigest ( )
print ( f " chktok: {{ chktok }} " )
print ( f " chkhsh: {{ chkhsh }} " )
res = None
# NOTE: if you get an error here, you need to update the convert-hf-to-gguf-update.py script
# or pull the latest version of the model from Huggingface
# don't edit the hashes manually!
{ src_ifs }
if res is None :
print ( " \\ n " )
print ( " ************************************************************************************** " )
print ( " ** WARNING: The BPE pre-tokenizer was not recognized! " )
print ( " ** There are 2 possible reasons for this: " )
print ( " ** - the model has not been added to convert-hf-to-gguf-update.py yet " )
print ( " ** - the pre-tokenization config has changed upstream " )
print ( " ** Check your model files and convert-hf-to-gguf-update.py and update them accordingly. " )
print ( " ** ref: https://github.com/ggerganov/llama.cpp/pull/6920 " )
print ( " ** " )
print ( f " ** chkhsh: {{ chkhsh }} " )
print ( " ************************************************************************************** " )
print ( " \\ n " )
raise NotImplementedError ( " BPE pre-tokenizer was not recognized - update get_vocab_base_pre() " )
print ( f " tokenizer.ggml.pre: {{ repr(res) }} " )
print ( f " chkhsh: {{ chkhsh }} " )
return res
"""
print ( src_func ) # noqa: NP100
logger . info ( " \n " )
logger . info ( " !!! Copy-paste the function above into convert-hf-to-gguf.py !!! " )
logger . info ( " \n " )
2024-04-29 13:58:41 +00:00
# generate tests for each tokenizer model
tests = [
" " ,
" " ,
" " ,
" " ,
" \t " ,
" \n " ,
" \n \n " ,
" \n \n \n " ,
" \t \n " ,
" Hello world " ,
" Hello world " ,
" Hello World " ,
" Hello World " ,
" Hello World! " ,
" Hello, world! " ,
" Hello, world! " ,
" this is 🦙.cpp " ,
" w048 7tuijk dsdfhu " ,
" нещо на Български " ,
" កាន់តែពិសេសអាចខលចេញ " ,
" 🚀 (normal) 😶🌫️ (multiple emojis concatenated) ✅ (only emoji that has its own token) " ,
" Hello " ,
" Hello " ,
" Hello " ,
" Hello " ,
" Hello " ,
" Hello \n Hello " ,
" ( " ,
" \n = " ,
" ' era " ,
" Hello, y ' all! How are you 😁 ?我想在apple工作1314151天~ " ,
" 3 " ,
" 33 " ,
" 333 " ,
" 3333 " ,
" 33333 " ,
" 333333 " ,
" 3333333 " ,
" 33333333 " ,
" 333333333 " ,
chktxt ,
]
# write the tests to ./models/ggml-vocab-{name}.gguf.inp
# the format is:
#
# test0
# __ggml_vocab_test__
# test1
# __ggml_vocab_test__
# ...
#
# with each model, encode all tests and write the results in ./models/ggml-vocab-{name}.gguf.out
# for each test, write the resulting tokens on a separate line
for model in models :
name = model [ " name " ]
tokt = model [ " tokt " ]
# create the tokenizer
tokenizer = AutoTokenizer . from_pretrained ( f " models/tokenizers/ { name } " )
2024-04-30 08:05:25 +00:00
with open ( f " models/ggml-vocab- { name } .gguf.inp " , " w " , encoding = " utf-8 " ) as f :
2024-04-29 13:58:41 +00:00
for text in tests :
f . write ( f " { text } " )
f . write ( " \n __ggml_vocab_test__ \n " )
with open ( f " models/ggml-vocab- { name } .gguf.out " , " w " ) as f :
for text in tests :
res = tokenizer . encode ( text , add_special_tokens = False )
for r in res :
f . write ( f " { r } " )
f . write ( " \n " )
2024-05-03 19:36:41 +00:00
logger . info ( f " Tests for { name } written in ./models/ggml-vocab- { name } .gguf.* " )
2024-04-29 13:58:41 +00:00
# generate commands for creating vocab files
2024-05-03 19:36:41 +00:00
logger . info ( " \n Run the following commands to generate the vocab files for testing: \n " )
2024-04-29 13:58:41 +00:00
for model in models :
name = model [ " name " ]
2024-05-03 19:36:41 +00:00
logger . info ( f " python3 convert-hf-to-gguf.py models/tokenizers/ { name } / --outfile models/ggml-vocab- { name } .gguf --vocab-only " )
2024-04-29 13:58:41 +00:00
2024-05-03 19:36:41 +00:00
logger . info ( " \n " )