mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-12 19:50:17 +00:00
28 lines
967 B
Markdown
28 lines
967 B
Markdown
# train-text-from-scratch
|
|
|
|
Basic usage instructions:
|
|
|
|
```bash
|
|
# get training data
|
|
wget https://raw.githubusercontent.com/brunoklein99/deep-learning-notes/master/shakespeare.txt
|
|
|
|
# train
|
|
./bin/llama-train-text-from-scratch \
|
|
--vocab-model ../models/ggml-vocab-llama.gguf \
|
|
--ctx 64 --embd 256 --head 8 --layer 16 \
|
|
--checkpoint-in chk-shakespeare-256x16-LATEST.gguf \
|
|
--checkpoint-out chk-shakespeare-256x16-ITERATION.gguf \
|
|
--model-out ggml-shakespeare-256x16-f32-ITERATION.gguf \
|
|
--train-data "shakespeare.txt" \
|
|
-t 6 -b 16 --seed 1 --adam-iter 256 \
|
|
--no-checkpointing
|
|
|
|
# predict
|
|
./bin/llama -m ggml-shakespeare-256x16-f32.gguf
|
|
```
|
|
|
|
Output files will be saved every N iterations (config with `--save-every N`).
|
|
The pattern "ITERATION" in the output filenames will be replaced with the iteration number and "LATEST" for the latest output.
|
|
|
|
To train GGUF models just pass them to `--checkpoint-in FN`.
|