This website requires JavaScript.
Explore
Help
Sign In
root
/
llama.cpp
Watch
1
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggerganov/llama.cpp.git
synced
2024-11-11 13:30:35 +00:00
Code
Issues
Actions
11
Packages
Projects
Releases
Wiki
Activity
All Workflows
build.yml
close-issue.yml
docker.yml
editorconfig.yml
gguf-publish.yml
labeler.yml
nix-ci-aarch64.yml
nix-ci.yml
nix-flake-update.yml
nix-publish-flake.yml
python-check-requirements.yml
python-lint.yml
python-type-check.yml
server.yml
Actor
All actors
root
Status
All status
success
failure
waiting
running
metal : more precise Q*K in FA vec kernel (#10247)
#1013
:
Commit
b0cefea58a
pushed by
root
master
2024-11-11 13:30:35 +00:00
0s
vulkan: Fix newly added tests for permuted mul_mat and 1D im2col (#10226)
#1001
:
Commit
160687b3ed
pushed by
root
master
2024-11-11 13:30:35 +00:00
0s
metal : reorder write loop in mul mat kernel + style (#10231)
#983
:
Commit
6423c65aa8
pushed by
root
master
2024-11-10 13:00:36 +00:00
0s
metal : opt-in compile flag for BF16 (#10218)
#972
:
Commit
ec450d3bbf
pushed by
root
master
2024-11-09 12:30:36 +00:00
0s
server : minor UI fix (#10207)
#964
:
Commit
76c6e7f105
pushed by
root
master
2024-11-08 20:10:36 +00:00
0s
ggml : add ggml-cpu.h to the public headers (#10204)
#956
:
Commit
97404c4a03
pushed by
root
master
2024-11-08 03:50:37 +00:00
0s
fix q4_0_8_8 format for corrupted tokens issue (#10198)
#948
:
Commit
2319126a70
pushed by
root
master
2024-11-07 19:40:37 +00:00
0s
ggml : adjust is_first_call init value (#10193)
#938
:
Commit
1dc04b2dee
pushed by
root
master
2024-11-07 11:30:37 +00:00
0s
llama : add <|tool_call|> formatting to Granite template (#10177)
#929
:
Commit
b8deef0ec0
pushed by
root
master
2024-11-06 11:00:36 +00:00
0s
ggml : fix arch check in bf16_to_fp32 (#10164)
#923
:
Commit
a9e8a9a030
pushed by
root
master
2024-11-05 18:40:37 +00:00
0s
ggml : fix q4xx mat mul, increase ggml_aligned_malloc alignment (#10167)
#909
:
Commit
401558b7ba
pushed by
root
master
2024-11-05 02:26:21 +00:00
0s
sync : ggml
#902
:
Commit
ce027adfb3
pushed by
root
master
2024-11-04 18:16:21 +00:00
0s
ggml : move CPU backend to a separate file (#10144)
#896
:
Commit
9f40989351
pushed by
root
master
2024-11-04 10:06:21 +00:00
0s
metal : minor fixup in FA kernel (#10143)
#889
:
Commit
08828a6d7d
pushed by
root
master
2024-11-04 01:56:20 +00:00
0s
server : fix slot selection by lru (#10126)
#866
:
Commit
42cadc74bd
pushed by
root
master
2024-11-03 17:46:20 +00:00
0s
llama : add simple-chat example (#10124)
#856
:
Commit
a6744e43e8
pushed by
root
master
2024-11-02 17:16:21 +00:00
0s
readme : update hot topics
#847
:
Commit
ba6f62eb79
pushed by
root
master
2024-11-02 00:56:21 +00:00
0s
sync : ggml
#838
:
Commit
815fe72adc
pushed by
root
master
2024-11-01 16:46:20 +00:00
0s
llama : improve output buffer type selection (#10098)
#832
:
Commit
85679d37f3
pushed by
root
master
2024-11-01 08:36:20 +00:00
0s
server : include scheme when printing URL (#10106)
#825
:
Commit
0a683e8088
pushed by
root
master
2024-11-01 00:26:20 +00:00
0s
kompute: add backend registry / device interfaces (#10045)
#816
:
Commit
61408e7fad
pushed by
root
master
2024-10-31 16:16:22 +00:00
0s
ggml : fix memory leaks when loading invalid gguf files (#10094)
#809
:
Commit
b9e02e8184
pushed by
root
master
2024-10-30 23:56:21 +00:00
0s
ggml : add Q4_0_8_8 RISC-V GEMV and GEMM kernels (#10029)
#802
:
Commit
fc83a9e584
pushed by
root
master
2024-10-30 15:46:20 +00:00
0s
ggml: Add POOL2D OP for GPU acceleration to the Vulkan backend in the MobileVLM model. (#9763)
#793
:
Commit
8f275a7c45
pushed by
root
master
2024-10-30 07:36:20 +00:00
0s
llama : Add IBM granite template (#10013)
#785
:
Commit
61715d5cc8
pushed by
root
master
2024-10-29 15:16:21 +00:00
0s
musa: workaround for Guilty Lockup in cleaning src0 (#10042)
#780
:
Commit
524afeec9d
pushed by
root
master
2024-10-28 22:56:21 +00:00
0s
llama : switch KQ multiplication to F32 precision by default (#10015)
#771
:
Commit
8841ce3f43
pushed by
root
master
2024-10-28 14:46:20 +00:00
0s
sync : ggml
#758
:
Commit
cc2983d375
pushed by
root
master
2024-10-27 15:36:17 +00:00
0s
llamafile : extend sgemm.cpp support for Q5_0 models (#10010)
#745
:
Commit
2f8bd2b901
pushed by
root
master
2024-10-26 13:46:20 +00:00
0s
server : refactor slot input data, move tokenizer to HTTP thread (#10023)
#733
:
Commit
958367bf53
pushed by
root
master
2024-10-25 13:16:21 +00:00
0s
First
Previous
1
2
3
4
Next
Last