mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2025-01-13 20:14:29 +00:00
cont : fix link
This commit is contained in:
parent
c172b322c2
commit
3e3cc7102f
@ -1161,7 +1161,7 @@ class FalconModel(Model):
|
||||
# So we rearrange them here,, so that we have n_head query weights
|
||||
# followed by n_head_kv key weights followed by n_head_kv value weights,
|
||||
# in contiguous fashion.
|
||||
# ref: https://github.com/jploski/ggml/blob/falcon40b/examples/falcon/convert-hf-to-gguf.py
|
||||
# ref: https://github.com/jploski/ggml/blob/falcon40b/examples/falcon/convert-hf-to-ggml.py
|
||||
|
||||
if "query_key_value" in name:
|
||||
n_head = self.find_hparam(["num_attention_heads", "n_head"])
|
||||
|
Loading…
Reference in New Issue
Block a user