Default Branch

master
Some checks are pending
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full-cuda.Dockerfile platforms:linux/amd64 tag:full-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/full.Dockerfile platforms:linux/amd64,linux/arm64 tag:full]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-cuda.Dockerfile platforms:linux/amd64 tag:light-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli-intel.Dockerfile platforms:linux/amd64 tag:light-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-cli.Dockerfile platforms:linux/amd64,linux/arm64 tag:light]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-cuda.Dockerfile platforms:linux/amd64 tag:server-cuda]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server-intel.Dockerfile platforms:linux/amd64 tag:server-intel]) (push) Waiting to run
Publish Docker image / Push Docker image to Docker Hub (map[dockerfile:.devops/llama-server.Dockerfile platforms:linux/amd64,linux/arm64 tag:server]) (push) Waiting to run
Nix CI / nix-eval (macos-latest) (push) Waiting to run
Nix CI / nix-eval (ubuntu-latest) (push) Waiting to run
Nix CI / nix-build (macos-latest) (push) Waiting to run
Nix CI / nix-build (ubuntu-latest) (push) Waiting to run
flake8 Lint / Lint (push) Waiting to run

a5b57b08ce · CUDA: enable Gemma FA for HIP/Pascal (#9581) · Updated 2024-09-22 07:34:52 +00:00

Branches

update_flake_lock_action
Some checks are pending
Python check requirements.txt / check-requirements (push) Waiting to run
flake8 Lint / Lint (push) Waiting to run
Python Type-Check / pyright type-check (push) Waiting to run

db660f5a40 · flake.lock: Update · Updated 2024-09-22 00:22:46 +00:00

3
1

6e873e561a · llama : make llm_tokenizer more private · Updated 2024-09-20 08:41:51 +00:00

14
2
gg/rerank
Some checks failed
flake8 Lint / Lint (push) Has been cancelled

5f95dccea8 · server : add rerank endpoint · Updated 2024-09-19 13:18:30 +00:00

31
7

ff231de553 · llama-bench : add time-to-first-byte stat · Updated 2024-09-19 06:15:29 +00:00

17
1

6b0248c29a · Update ggml/src/ggml.c · Updated 2024-09-18 16:00:26 +00:00

19
2
pr_add_intel_amx_support
Some checks failed
Python check requirements.txt / check-requirements (push) Has been cancelled
flake8 Lint / Lint (push) Has been cancelled
Python Type-Check / pyright type-check (push) Has been cancelled

c90a43a237 · minor change · Updated 2024-09-18 07:31:08 +00:00

22
13
fix_ctx_default
Some checks failed
flake8 Lint / Lint (push) Has been cancelled

a6a8f8d09c · Update docs/backend/SYCL.md · Updated 2024-09-17 08:25:43 +00:00

51
2

695c4483a0 · wip · Updated 2024-09-16 21:28:36 +00:00

30
1
gg/cb-naming
Some checks failed
Python check requirements.txt / check-requirements (push) Has been cancelled
flake8 Lint / Lint (push) Has been cancelled
Python Type-Check / pyright type-check (push) Has been cancelled

cc1c017191 · naming : normalize the name of callback-related identifiers · Updated 2024-09-16 06:11:42 +00:00

39
1

73ef3f769c · Update llama-server-intel.Dockerfile · Updated 2024-09-15 15:21:46 +00:00

44
3
gg/cmake-dedup-link
Some checks failed
flake8 Lint / Lint (push) Has been cancelled

fb8f142554 · one more CMAKE_CXX_FLAGS fix (#9471) · Updated 2024-09-13 13:13:07 +00:00

53
5

d7c042d1ae · ggml : make n_threads_cur atomic_int · Updated 2024-09-11 18:12:11 +00:00

69
1

f9968f661d · ggml : update comments [no ci] · Updated 2024-09-11 10:16:39 +00:00

82
5

2d79a7077c · quantize : use unused imatrix chunk_size with LLAMA_TRACE · Updated 2024-09-10 16:09:17 +00:00    root

99
13

cfbf33a705 · ggml : style changes + fix 512-bit nb loop check · Updated 2024-09-09 09:50:35 +00:00    root

143
4

c3e2bb6dcf · rpc : fix nkvo · Updated 2024-09-07 01:24:47 +00:00    root

124
1

b979fc97ba · cmake : use ggml-metal.metal from source dir to build default.metallib · Updated 2024-09-05 16:17:56 +00:00    root

133
1

75b3a09602 · test-backend-ops : add TQ1_0 and TQ2_0 comments for later · Updated 2024-09-04 19:00:21 +00:00    root

135
33

f648ca2cee · llama : add llama_sampling API + move grammar in libllama · Updated 2024-09-03 07:31:54 +00:00    root

142
1

40fa68cb46 · readme : add API change notice · Updated 2024-09-02 15:32:24 +00:00    root

151
3