mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-11-11 21:39:52 +00:00
1c641e6aac
* `main`/`server`: rename to `llama` / `llama-server` for consistency w/ homebrew
* server: update refs -> llama-server
gitignore llama-server
* server: simplify nix package
* main: update refs -> llama
fix examples/main ref
* main/server: fix targets
* update more names
* Update build.yml
* rm accidentally checked in bins
* update straggling refs
* Update .gitignore
* Update server-llm.sh
* main: target name -> llama-cli
* Prefix all example bins w/ llama-
* fix main refs
* rename {main->llama}-cmake-pkg binary
* prefix more cmake targets w/ llama-
* add/fix gbnf-validator subfolder to cmake
* sort cmake example subdirs
* rm bin files
* fix llama-lookup-* Makefile rules
* gitignore /llama-*
* rename Dockerfiles
* rename llama|main -> llama-cli; consistent RPM bin prefixes
* fix some missing -cli suffixes
* rename dockerfile w/ llama-cli
* rename(make): llama-baby-llama
* update dockerfile refs
* more llama-cli(.exe)
* fix test-eval-callback
* rename: llama-cli-cmake-pkg(.exe)
* address gbnf-validator unused fread warning (switched to C++ / ifstream)
* add two missing llama- prefixes
* Updating docs for eval-callback binary to use new `llama-` prefix.
* Updating a few lingering doc references for rename of main to llama-cli
* Updating `run-with-preset.py` to use new binary names.
Updating docs around `perplexity` binary rename.
* Updating documentation references for lookup-merge and export-lora
* Updating two small `main` references missed earlier in the finetune docs.
* Update apps.nix
* update grammar/README.md w/ new llama-* names
* update llama-rpc-server bin name + doc
* Revert "update llama-rpc-server bin name + doc"
This reverts commit e474ef1df4
.
* add hot topic notice to README.md
* Update README.md
* Update README.md
* rename gguf-split & quantize bins refs in **/tests.sh
---------
Co-authored-by: HanClinto <hanclinto@gmail.com>
152 lines
4.9 KiB
Bash
Executable File
152 lines
4.9 KiB
Bash
Executable File
#!/bin/bash
|
|
|
|
set -euo pipefail
|
|
|
|
cd "$(dirname "$0")/.." || exit
|
|
|
|
if [[ -z "${PROMPT_CACHE_FILE+x}" || -z "${CHAT_SAVE_DIR+x}" ]]; then
|
|
echo >&2 "error: PROMPT_CACHE_FILE and CHAT_SAVE_DIR must be provided"
|
|
exit 1
|
|
fi
|
|
|
|
MODEL="${MODEL:-./models/llama-13b/ggml-model-q4_0.gguf}"
|
|
PROMPT_TEMPLATE="${PROMPT_TEMPLATE:-./prompts/chat.txt}"
|
|
USER_NAME="${USER_NAME:-User}"
|
|
AI_NAME="${AI_NAME:-ChatLLaMa}"
|
|
DATE_TIME="$(date +%H:%M)"
|
|
DATE_YEAR="$(date +%Y)"
|
|
|
|
LOG="${CHAT_SAVE_DIR}/main.log"
|
|
LOG_BG="${CHAT_SAVE_DIR}/main-bg.log"
|
|
CUR_PROMPT_FILE="${CHAT_SAVE_DIR}/current-prompt.txt"
|
|
CUR_PROMPT_CACHE="${CHAT_SAVE_DIR}/current-cache.bin"
|
|
NEXT_PROMPT_FILE="${CHAT_SAVE_DIR}/next-prompt.txt"
|
|
NEXT_PROMPT_CACHE="${CHAT_SAVE_DIR}/next-cache.bin"
|
|
|
|
SESSION_SIZE_MSG_PATTERN='main: session file matches [[:digit:]]+ / [[:digit:]]+'
|
|
SAMPLE_TIME_MSG_PATTERN='sample time =[[:space:]]+[[:digit:]]+.[[:digit:]]+ ms /[[:space:]]+[[:digit:]]+'
|
|
SED_DELETE_MESSAGES="/^(${USER_NAME}:|${AI_NAME}:|\\.\\.\\.)/,\$d"
|
|
|
|
CTX_SIZE=2048
|
|
CTX_ROTATE_POINT=$((CTX_SIZE * 3 / 5)) # REVIEW
|
|
OPTS=(--model "$MODEL" --ctx_size "$CTX_SIZE" --repeat_last_n 256 "$@")
|
|
|
|
# An unbuffered `tail -c+N`
|
|
skip_bytes() {
|
|
LANG=C IFS= read -r -n "$1" -d '' c
|
|
while LANG=C IFS= read -r -n 1 -d '' c; do
|
|
printf '%s' "$c"
|
|
done
|
|
}
|
|
|
|
mkdir -p "$CHAT_SAVE_DIR"
|
|
echo >"$LOG"
|
|
trap "tail -n100 ${LOG}" EXIT
|
|
|
|
if [[ ! -e "$CUR_PROMPT_FILE" ]]; then
|
|
sed -e "s/\[\[USER_NAME\]\]/${USER_NAME}/g" \
|
|
-e "s/\[\[AI_NAME\]\]/${AI_NAME}/g" \
|
|
-e "s/\[\[DATE_TIME\]\]/${DATE_TIME}/g" \
|
|
-e "s/\[\[DATE_YEAR\]\]/${DATE_YEAR}/g" \
|
|
"$PROMPT_TEMPLATE" >"$CUR_PROMPT_FILE"
|
|
fi
|
|
|
|
if [[ ! -e "$NEXT_PROMPT_FILE" ]]; then
|
|
sed -r "$SED_DELETE_MESSAGES" "$CUR_PROMPT_FILE" >"$NEXT_PROMPT_FILE"
|
|
fi
|
|
|
|
if [[ "$(tail -c4 "$NEXT_PROMPT_FILE")" != "..." ]]; then
|
|
echo '...' >>"$NEXT_PROMPT_FILE"
|
|
fi
|
|
|
|
if [[ ! -e "$PROMPT_CACHE_FILE" ]]; then
|
|
echo 'Prompt cache does not exist, building...'
|
|
# Default batch_size to 64 here for better user feedback during initial prompt processing
|
|
./llama-cli 2>>"$LOG" \
|
|
--batch_size 64 \
|
|
"${OPTS[@]}" \
|
|
--prompt-cache "$PROMPT_CACHE_FILE" \
|
|
--file "$CUR_PROMPT_FILE" \
|
|
--n_predict 1
|
|
echo
|
|
echo 'Done!'
|
|
fi
|
|
|
|
if [[ ! -e "$CUR_PROMPT_CACHE" ]]; then
|
|
cp "$PROMPT_CACHE_FILE" "$CUR_PROMPT_CACHE"
|
|
fi
|
|
if [[ ! -e "$NEXT_PROMPT_CACHE" ]]; then
|
|
cp "$PROMPT_CACHE_FILE" "$NEXT_PROMPT_CACHE"
|
|
fi
|
|
|
|
printf '%s ' "$(< "$CUR_PROMPT_FILE")"
|
|
n_tokens=0
|
|
|
|
while read -e line; do
|
|
# Limit generation to remaining context, with a buffer and estimating 2 chars/token for input
|
|
n_predict=$((CTX_SIZE - n_tokens - ${#line} / 2 - 32))
|
|
|
|
# Swap prompts when we're about to run out of context
|
|
if ((n_predict <= 0)); then
|
|
wait # for background main (below) to finish with next prompt
|
|
mv "$NEXT_PROMPT_FILE" "$CUR_PROMPT_FILE"
|
|
mv "$NEXT_PROMPT_CACHE" "$CUR_PROMPT_CACHE"
|
|
|
|
sed -r "$SED_DELETE_MESSAGES" "$CUR_PROMPT_FILE" >"$NEXT_PROMPT_FILE"
|
|
echo '...' >>"$NEXT_PROMPT_FILE"
|
|
cp "$PROMPT_CACHE_FILE" "$NEXT_PROMPT_CACHE"
|
|
|
|
n_tokens=0
|
|
n_predict=$((CTX_SIZE / 2))
|
|
fi
|
|
|
|
echo " ${line}" >>"$CUR_PROMPT_FILE"
|
|
if ((n_tokens > CTX_ROTATE_POINT)); then
|
|
echo " ${line}" >>"$NEXT_PROMPT_FILE"
|
|
fi
|
|
|
|
n_prompt_len_pre=$(($(wc -c <"$CUR_PROMPT_FILE")))
|
|
|
|
printf '%s: ' "$AI_NAME" >>"$CUR_PROMPT_FILE"
|
|
|
|
./llama-cli 2>>"$LOG" "${OPTS[@]}" \
|
|
--prompt-cache "$CUR_PROMPT_CACHE" \
|
|
--prompt-cache-all \
|
|
--file "$CUR_PROMPT_FILE" \
|
|
--reverse-prompt "${USER_NAME}:" \
|
|
--n_predict "$n_predict" |
|
|
skip_bytes 1 | # skip BOS token added by ./llama-cli
|
|
tee "$CUR_PROMPT_FILE.tmp" | # save prompt + generation to tmp file
|
|
skip_bytes "$n_prompt_len_pre" # print generation
|
|
|
|
mv "$CUR_PROMPT_FILE.tmp" "$CUR_PROMPT_FILE"
|
|
|
|
# if we hit n_predict instead of reverse-prompt, we need to add the prompt
|
|
if [[ "$(tail -n1 "$CUR_PROMPT_FILE")" != "${USER_NAME}:" ]]; then
|
|
printf '\n%s:' "$USER_NAME"
|
|
printf '\n%s:' "$USER_NAME" >> "$CUR_PROMPT_FILE"
|
|
fi
|
|
|
|
printf ' '
|
|
|
|
# HACK get num tokens from debug message
|
|
# TODO get both messages in one go
|
|
if ! session_size_msg="$(tail -n30 "$LOG" | grep -oE "$SESSION_SIZE_MSG_PATTERN")" ||
|
|
! sample_time_msg="$(tail -n10 "$LOG" | grep -oE "$SAMPLE_TIME_MSG_PATTERN")"; then
|
|
echo >&2 "Couldn't get number of tokens from ./llama-cli output!"
|
|
exit 1
|
|
fi
|
|
|
|
n_tokens=$(($(cut -d/ -f2 <<<"$session_size_msg") + $(cut -d/ -f2 <<<"$sample_time_msg")))
|
|
|
|
if ((n_tokens > CTX_ROTATE_POINT)); then
|
|
tail -c+$((n_prompt_len_pre + 1)) "$CUR_PROMPT_FILE" >>"$NEXT_PROMPT_FILE"
|
|
fi
|
|
|
|
# Update cache for next prompt in background, ideally during user input
|
|
./llama-cli >>"$LOG_BG" 2>&1 "${OPTS[@]}" \
|
|
--prompt-cache "$NEXT_PROMPT_CACHE" \
|
|
--file "$NEXT_PROMPT_FILE" \
|
|
--n_predict 1 &
|
|
done
|