This website requires JavaScript.
Explore
Help
Sign In
root
/
llama.cpp
Watch
1
Star
0
Fork
0
You've already forked llama.cpp
mirror of
https://github.com/ggerganov/llama.cpp.git
synced
2025-01-12 19:50:17 +00:00
Code
Issues
Actions
5
Packages
Projects
Releases
Wiki
Activity
All Workflows
build.yml
close-issue.yml
docker.yml
editorconfig.yml
gguf-publish.yml
labeler.yml
python-check-requirements.yml
python-lint.yml
python-type-check.yml
server.yml
Actor
All actors
root
Status
All status
success
failure
waiting
running
gguf-py: fixed local detection of gguf package (#11180)
#1682
:
Commit
c05e8c9934
pushed by
root
master
2025-01-12 11:40:17 +00:00
0s
convert : add --print-supported-models option (#11172)
#1675
:
Commit
ff3fcabc72
pushed by
root
master
2025-01-11 11:11:46 +00:00
0s
llama: add support for QRWKV6 model architecture (#11001)
#1667
:
Commit
ee7136c6d1
pushed by
root
master
2025-01-10 10:41:48 +00:00
0s
doc: add cuda guide for fedora (#11135)
#1661
:
Commit
1204f97270
pushed by
root
master
2025-01-10 02:31:46 +00:00
0s
fix: add missing msg in static_assert (#11143)
#1651
:
Commit
8d59d91171
pushed by
root
master
2025-01-09 18:21:45 +00:00
0s
ci : use actions from ggml-org (#11140)
#1646
:
Commit
f7cd13301c
pushed by
root
master
2025-01-09 02:01:45 +00:00
0s
llama : remove unused headers (#11109)
#1638
:
Commit
ecebbd292d
pushed by
root
master
2025-01-07 17:21:42 +00:00
0s
Vulkan: Add device-specific blacklist for coopmat for the AMD proprietary driver (#11074)
#1628
:
Commit
b56f079e28
pushed by
root
master
2025-01-06 02:44:32 +00:00
0s
llama : add support for the cohere2 model architecture (#10900)
#1623
:
Commit
46be942214
pushed by
root
master
2025-01-05 00:04:37 +00:00
0s
server: bench: minor fixes (#10765)
#1615
:
Commit
2f0ee84b9b
pushed by
root
master
2025-01-04 02:44:32 +00:00
0s
server : allow using LoRA adapters per-request (#10994)
#1609
:
Commit
0da5d86026
pushed by
root
master
2025-01-02 23:04:35 +00:00
0s
ggml : fixes for AVXVNNI instruction set with MSVC and Clang (#11027)
#1599
:
Commit
0827b2c1da
pushed by
root
master
2025-01-02 02:44:32 +00:00
0s
server : add OAI compat for /v1/completions (#10974)
#1596
:
Commit
5896c65232
pushed by
root
master
2024-12-31 22:04:36 +00:00
0s
server : add support for "encoding_format": "base64" to the */embeddings endpoints (#10967)
#1576
:
Commit
9ba399dfa7
pushed by
root
master
2024-12-26 08:44:32 +00:00
0s
ggml : more perfo with llamafile tinyblas on x86_64 (#10714)
#1569
:
Commit
2cd43f4900
pushed by
root
master
2024-12-25 02:44:36 +00:00
0s
server : fix missing model id in /model endpoint (#10957)
#1564
:
Commit
14b699ecde
pushed by
root
master
2024-12-24 18:34:36 +00:00
0s
llama : support InfiniAI Megrez 3b (#10893)
#1553
:
Commit
b92a14a841
pushed by
root
master
2024-12-23 18:04:35 +00:00
0s
convert : add BertForMaskedLM (#10919)
#1547
:
Commit
5cd85b5e00
pushed by
root
master
2024-12-22 14:44:32 +00:00
0s
SYCL: Migrate away from deprecated ggml_tensor->backend (#10840)
#1540
:
Commit
eb5c3dc64b
pushed by
root
master
2024-12-21 08:54:36 +00:00
0s
llama : minor grammar refactor (#10897)
#1529
:
Commit
5cab3e4aaa
pushed by
root
master
2024-12-20 16:34:36 +00:00
0s
ggml : fix arm build (#10890)
#1524
:
Commit
9177484f58
pushed by
root
master
2024-12-19 16:04:37 +00:00
0s
server: avoid overwriting Authorization header (#10878)
#1519
:
Commit
7bbb5acf12
pushed by
root
master
2024-12-18 23:44:35 +00:00
0s
Revert "llama : add Falcon3 support (#10864)" (#10876)
#1514
:
Commit
4da69d1abd
pushed by
root
master
2024-12-18 15:34:37 +00:00
0s
Use model->gguf_kv for loading the template instead of using the C API. (#10868)
#1504
:
Commit
d62b532c52
pushed by
root
master
2024-12-18 07:24:36 +00:00
0s
llava : Allow locally downloaded models for QwenVL (#10833)
#1490
:
Commit
4ddd199f6f
pushed by
root
master
2024-12-17 02:44:32 +00:00
0s
server: Fix `has_next_line` in JSON response (#10818)
#1483
:
Commit
89d604f2c8
pushed by
root
master
2024-12-15 22:14:36 +00:00
0s
llama : add Qwen2VL support + multimodal RoPE (#10361)
#1477
:
Commit
ba1cb19cdd
pushed by
root
master
2024-12-15 05:54:36 +00:00
0s
Introducing experimental OpenCL backend with support for Qualcomm Adreno GPUs (#10693)
#1471
:
Commit
a76c56fa1a
pushed by
root
master
2024-12-14 13:34:35 +00:00
0s
ci : build docker images only once daily (#10503)
#1467
:
Commit
50d5cecbda
pushed by
root
b4172
2024-12-13 14:09:49 +00:00
0s
ggml : optimize Q4_0 into Q4_0_X_Y repack (#10324)
#1466
:
Commit
1e58ee1318
pushed by
root
b4096
2024-12-13 14:09:49 +00:00
0s
First
Previous
1
2
3
4
5
...
Next
Last