Georgi Gerganov
|
f00780b2ee
|
llama : sync gguf-llama.cpp with latest llama.cpp (#2608)
* llama : sync gguf-llama.cpp with latest llama.cpp
* minor : indentation + assert
* llama : refactor gguf_buffer and gguf_ctx_buffer
* llama : minor
|
2023-08-14 16:28:44 +03:00 |
|
Georgi Gerganov
|
0c19ae70d5
|
simple : minor style changes
|
2023-08-14 12:58:12 +03:00 |
|
M. Yusuf Sarıgöz
|
202eab04d3
|
gguf : quantization is working
|
2023-08-12 16:39:05 +03:00 |
|
M. Yusuf Sarıgöz
|
1fc3d30b71
|
gguf : start implementing quantization (WIP)
|
2023-08-12 16:09:47 +03:00 |
|
M. Yusuf Sarıgöz
|
b2571af255
|
gguf : start implementing quantization (WIP)
|
2023-08-12 14:28:17 +03:00 |
|
M. Yusuf Sarıgöz
|
1c4d8bf981
|
gguf : start implementing libllama in GGUF (WIP)
|
2023-08-10 16:52:08 +03:00 |
|
Georgi Gerganov
|
8083ae347a
|
gguf : minor stuff
|
2023-08-07 19:02:18 +03:00 |
|
M. Yusuf Sarıgöz
|
cf365fbc20
|
gguf : gguf counterpart of llama-util.h
|
2023-08-02 11:13:56 +03:00 |
|
M. Yusuf Sarıgöz
|
bb42aefaeb
|
gguf : mmap tensor data example
|
2023-07-31 17:46:12 +03:00 |
|
M. Yusuf Sarıgöz
|
08dc8fd884
|
gguf : do not hardcode tensor names to read
|
2023-07-29 10:24:46 +03:00 |
|
klosax
|
3492f848d7
|
gguf : add gguf_find_key (#2438)
* gguf.cpp : find key example
* ggml.h : add gguf_find_key
* ggml.c : add gguf_find_key
|
2023-07-28 23:45:24 +03:00 |
|
Georgi Gerganov
|
d2b6ca13ad
|
gguf : add array support
|
2023-07-27 14:53:07 +03:00 |
|
Georgi Gerganov
|
d89533dff6
|
gguf : expose the gguf_type enum through the API for now
|
2023-07-27 11:10:34 +03:00 |
|
Georgi Gerganov
|
5628ec7163
|
gguf : read / write sample models
|
2023-07-26 22:40:45 +03:00 |
|
Georgi Gerganov
|
4d698495ea
|
gguf : init
|
2023-07-26 18:21:12 +03:00 |
|