• Joined on 2024-09-10
root synced commits to refs/tags/b3801 at root/llama.cpp from mirror 2024-09-22 04:56:21 +00:00
root synced new reference refs/tags/b3801 to root/llama.cpp from mirror 2024-09-22 04:56:21 +00:00
root synced and deleted reference refs/tags/refs/pull/9454/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
root synced and deleted reference refs/tags/refs/pull/9574/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
root synced commits to master at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
ecd5d6b65b llama: remove redundant loop when constructing ubatch (#9574)
2a63caaa69 RWKV v6: RWKV_WKV op CUDA implementation (#9454)
Compare 2 commits »
root synced commits to update_flake_lock_action at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
db660f5a40 flake.lock: Update
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
63351143b2 quantize : improve type name parsing (#9570)
Compare 45 commits »
root synced commits to refs/pull/8208/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
63351143b2 quantize : improve type name parsing (#9570)
Compare 15 commits »
root synced commits to refs/pull/9058/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
ecd5d6b65b llama: remove redundant loop when constructing ubatch (#9574)
2a63caaa69 RWKV v6: RWKV_WKV op CUDA implementation (#9454)
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 4 commits »
root synced commits to refs/pull/9096/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
ecd5d6b65b llama: remove redundant loop when constructing ubatch (#9574)
2a63caaa69 RWKV v6: RWKV_WKV op CUDA implementation (#9454)
Compare 3 commits »
root synced commits to refs/pull/9186/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 2 commits »
root synced commits to refs/pull/9325/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 2 commits »
root synced commits to refs/pull/9438/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
ecd5d6b65b llama: remove redundant loop when constructing ubatch (#9574)
2a63caaa69 RWKV v6: RWKV_WKV op CUDA implementation (#9454)
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 4 commits »
root synced commits to refs/pull/9454/head at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
caeba159da Merge branch 'master' into wkv-cuda
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
63351143b2 quantize : improve type name parsing (#9570)
Compare 52 commits »
root synced commits to refs/pull/9482/merge at root/llama.cpp from mirror 2024-09-22 04:56:20 +00:00
ecd5d6b65b llama: remove redundant loop when constructing ubatch (#9574)
2a63caaa69 RWKV v6: RWKV_WKV op CUDA implementation (#9454)
Compare 3 commits »
root synced and deleted reference refs/tags/sycl-fallback-mmvq at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
root synced commits to refs/pull/8837/merge at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
Compare 4 commits »
root synced commits to refs/pull/8878/merge at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
Compare 4 commits »
root synced commits to refs/pull/9096/merge at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 2 commits »
root synced commits to refs/pull/9209/merge at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
Compare 2 commits »
root synced commits to refs/pull/9217/merge at root/llama.cpp from mirror 2024-09-21 20:46:21 +00:00
d09770cae7 ggml-alloc : fix list of allocated tensors with GGML_ALLOCATOR_DEBUG (#9573)
41f477879f Update CUDA graph on scale change plus clear nodes/params (#9550)
e948a7da7a CI: Provide prebuilt windows binary for hip (#9467)
63351143b2 quantize : improve type name parsing (#9570)
Compare 17 commits »