mirror of
https://github.com/ggerganov/llama.cpp.git
synced 2024-12-30 21:34:36 +00:00
convert-llama-h5-to-gguf.py : simplify nbytes
This commit is contained in:
parent
f8218477b3
commit
e606ffeaee
@ -196,17 +196,7 @@ for name in list_vars.keys():
|
|||||||
# if f16 desired, convert any float32 2-dim weight tensors to float16
|
# if f16 desired, convert any float32 2-dim weight tensors to float16
|
||||||
data_dtype = np.float16
|
data_dtype = np.float16
|
||||||
|
|
||||||
nelements = 1
|
data_nbytes = data.size * 2 if data_dtype == np.float16 else data.size * 4
|
||||||
|
|
||||||
for i in range(n_dims):
|
|
||||||
nelements *= data.shape[n_dims - 1 - i]
|
|
||||||
|
|
||||||
data_nbytes = 0
|
|
||||||
if data_dtype == np.float16:
|
|
||||||
data_nbytes = nelements * 2
|
|
||||||
elif data_dtype == np.float32:
|
|
||||||
data_nbytes = nelements * 4
|
|
||||||
|
|
||||||
|
|
||||||
gguf_writer.add_tensor_info(name, data.shape, data_dtype, data_nbytes)
|
gguf_writer.add_tensor_info(name, data.shape, data_dtype, data_nbytes)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user