#!/bin/bash set -euo pipefail cd "$(dirname "$0")/.." || exit if [[ -z "${PROMPT_CACHE_FILE+x}" || -z "${CHAT_SAVE_DIR+x}" ]]; then echo >&2 "error: PROMPT_CACHE_FILE and CHAT_SAVE_DIR must be provided" exit 1 fi MODEL="${MODEL:-./models/llama-13b/ggml-model-q4_0.gguf}" PROMPT_TEMPLATE="${PROMPT_TEMPLATE:-./prompts/chat.txt}" USER_NAME="${USER_NAME:-User}" AI_NAME="${AI_NAME:-ChatLLaMa}" DATE_TIME="$(date +%H:%M)" DATE_YEAR="$(date +%Y)" LOG="${CHAT_SAVE_DIR}/main.log" LOG_BG="${CHAT_SAVE_DIR}/main-bg.log" CUR_PROMPT_FILE="${CHAT_SAVE_DIR}/current-prompt.txt" CUR_PROMPT_CACHE="${CHAT_SAVE_DIR}/current-cache.bin" NEXT_PROMPT_FILE="${CHAT_SAVE_DIR}/next-prompt.txt" NEXT_PROMPT_CACHE="${CHAT_SAVE_DIR}/next-cache.bin" SESSION_AND_SAMPLE_PATTERN='main: session file matches [[:digit:]]+ / [[:digit:]]+'\ '|'\ 'sampling time =[[:space:]]+[[:digit:]]+.[[:digit:]]+ ms /[[:space:]]+[[:digit:]]+' SED_DELETE_MESSAGES="/^(${USER_NAME}:|${AI_NAME}:|\\.\\.\\.)/,\$d" CTX_SIZE=2048 CTX_ROTATE_POINT=$((CTX_SIZE * 3 / 5)) # REVIEW OPTS=(--model "$MODEL" --ctx_size "$CTX_SIZE" --repeat_last_n 256 "$@") # An unbuffered `tail -c+N` skip_bytes() { LANG=C IFS= read -r -n "$1" -d '' c while LANG=C IFS= read -r -n 1 -d '' c; do printf '%s' "$c" done } mkdir -p "$CHAT_SAVE_DIR" echo >"$LOG" trap "tail -n100 ${LOG}" EXIT if [[ ! -e "$CUR_PROMPT_FILE" ]]; then sed -e "s/\[\[USER_NAME\]\]/${USER_NAME}/g" \ -e "s/\[\[AI_NAME\]\]/${AI_NAME}/g" \ -e "s/\[\[DATE_TIME\]\]/${DATE_TIME}/g" \ -e "s/\[\[DATE_YEAR\]\]/${DATE_YEAR}/g" \ "$PROMPT_TEMPLATE" >"$CUR_PROMPT_FILE" fi if [[ ! -e "$NEXT_PROMPT_FILE" ]]; then sed -r "$SED_DELETE_MESSAGES" "$CUR_PROMPT_FILE" >"$NEXT_PROMPT_FILE" fi if [[ "$(tail -c4 "$NEXT_PROMPT_FILE")" != "..." ]]; then echo '...' >>"$NEXT_PROMPT_FILE" fi if [[ ! -e "$PROMPT_CACHE_FILE" ]]; then echo 'Prompt cache does not exist, building...' # Default batch_size to 64 here for better user feedback during initial prompt processing ./llama-cli 2>>"$LOG" \ --batch_size 64 \ "${OPTS[@]}" \ --prompt-cache "$PROMPT_CACHE_FILE" \ --file "$CUR_PROMPT_FILE" \ --n_predict 1 echo echo 'Done!' fi if [[ ! -e "$CUR_PROMPT_CACHE" ]]; then cp "$PROMPT_CACHE_FILE" "$CUR_PROMPT_CACHE" fi if [[ ! -e "$NEXT_PROMPT_CACHE" ]]; then cp "$PROMPT_CACHE_FILE" "$NEXT_PROMPT_CACHE" fi printf '%s ' "$(< "$CUR_PROMPT_FILE")" n_tokens=0 while read -e line; do # Limit generation to remaining context, with a buffer and estimating 2 chars/token for input n_predict=$((CTX_SIZE - n_tokens - ${#line} / 2 - 32)) # Swap prompts when we're about to run out of context if ((n_predict <= 0)); then wait # for background main (below) to finish with next prompt mv "$NEXT_PROMPT_FILE" "$CUR_PROMPT_FILE" mv "$NEXT_PROMPT_CACHE" "$CUR_PROMPT_CACHE" sed -r "$SED_DELETE_MESSAGES" "$CUR_PROMPT_FILE" >"$NEXT_PROMPT_FILE" echo '...' >>"$NEXT_PROMPT_FILE" cp "$PROMPT_CACHE_FILE" "$NEXT_PROMPT_CACHE" n_tokens=0 n_predict=$((CTX_SIZE / 2)) fi echo " ${line}" >>"$CUR_PROMPT_FILE" if ((n_tokens > CTX_ROTATE_POINT)); then echo " ${line}" >>"$NEXT_PROMPT_FILE" fi n_prompt_len_pre=$(($(wc -c <"$CUR_PROMPT_FILE"))) printf '%s: ' "$AI_NAME" >>"$CUR_PROMPT_FILE" ./llama-cli 2>>"$LOG" "${OPTS[@]}" \ --prompt-cache "$CUR_PROMPT_CACHE" \ --prompt-cache-all \ --file "$CUR_PROMPT_FILE" \ --reverse-prompt "${USER_NAME}:" \ --n_predict "$n_predict" | skip_bytes 1 | # skip BOS token added by ./llama-cli tee "$CUR_PROMPT_FILE.tmp" | # save prompt + generation to tmp file skip_bytes "$n_prompt_len_pre" # print generation mv "$CUR_PROMPT_FILE.tmp" "$CUR_PROMPT_FILE" # if we hit n_predict instead of reverse-prompt, we need to add the prompt if [[ "$(tail -n1 "$CUR_PROMPT_FILE")" != "${USER_NAME}:" ]]; then printf '\n%s:' "$USER_NAME" printf '\n%s:' "$USER_NAME" >> "$CUR_PROMPT_FILE" fi printf ' ' if ! session_and_sample_msg=$(tail -n30 "$LOG" | grep -oE "$SESSION_AND_SAMPLE_PATTERN"); then echo >&2 "Couldn't get number of tokens from ./llama-cli output!" exit 1 fi n_tokens=$(awk '{sum+=$1} END {print sum}' <<< "$(cut -d/ -f2 <<< "$session_and_sample_msg")") if ((n_tokens > CTX_ROTATE_POINT)); then tail -c+$((n_prompt_len_pre + 1)) "$CUR_PROMPT_FILE" >>"$NEXT_PROMPT_FILE" fi # Update cache for next prompt in background, ideally during user input ./llama-cli >>"$LOG_BG" 2>&1 "${OPTS[@]}" \ --prompt-cache "$NEXT_PROMPT_CACHE" \ --file "$NEXT_PROMPT_FILE" \ --n_predict 1 & done