2023-11-01 11:49:04 +00:00
#!/bin/bash
cd ` dirname $0 `
cd ../..
2024-06-12 23:41:52 +00:00
EXE = "./llama-finetune"
2023-11-01 11:49:04 +00:00
if [ [ ! $LLAMA_MODEL_DIR ] ] ; then LLAMA_MODEL_DIR = "./models" ; fi
if [ [ ! $LLAMA_TRAINING_DIR ] ] ; then LLAMA_TRAINING_DIR = "." ; fi
# MODEL="$LLAMA_MODEL_DIR/openllama-3b-v2-q8_0.gguf" # This is the model the readme uses.
2024-07-07 10:37:47 +00:00
MODEL = " $LLAMA_MODEL_DIR /openllama-3b-v2.gguf " # An f16 model. Note in this case with "-g", you get an f32-format .BIN file that isn't yet supported if you use it with "llama-cli --lora" with GPU inferencing.
2023-11-01 11:49:04 +00:00
while getopts "dg" opt; do
case $opt in
d)
DEBUGGER = "gdb --args"
; ;
g)
EXE = "./build/bin/Release/finetune"
GPUARG = "--gpu-layers 25"
; ;
esac
done
$DEBUGGER $EXE \
--model-base $MODEL \
$GPUARG \
--checkpoint-in chk-ol3b-shakespeare-LATEST.gguf \
--checkpoint-out chk-ol3b-shakespeare-ITERATION.gguf \
--lora-out lora-ol3b-shakespeare-ITERATION.bin \
--train-data " $LLAMA_TRAINING_DIR \shakespeare.txt " \
--save-every 10 \
--threads 10 --adam-iter 30 --batch 4 --ctx 64 \
--use-checkpointing