Skip to content

Actions: katsu560/llama.cpp

Publish Docker image

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
22 workflow runs
22 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

ggml : do not use ARM features not included in the build (#10457)
Publish Docker image #62: Commit 55ed008 pushed by katsu560
November 23, 2024 16:13 31m 53s master
November 23, 2024 16:13 31m 53s
gguf-py : fix double call to add_architecture() (#8952)
Publish Docker image #61: Commit 911b437 pushed by katsu560
August 10, 2024 07:02 8m 29s master
August 10, 2024 07:02 8m 29s
batched-bench : handle empty -npl (#8839)
Publish Docker image #60: Commit ecf6b7f pushed by katsu560
August 4, 2024 11:55 24m 17s master
August 4, 2024 11:55 24m 17s
Vulkan MMQ Fix (#8479)
Publish Docker image #59: Commit bda62d7 pushed by katsu560
July 15, 2024 09:40 7m 52s master
July 15, 2024 09:40 7m 52s
llama : return nullptr from llama_grammar_init (#8093)
Publish Docker image #58: Commit e6bf007 pushed by katsu560
June 25, 2024 19:10 9m 4s master
June 25, 2024 19:10 9m 4s
server : new UI (#7633)
Publish Docker image #57: Commit 2e66683 pushed by katsu560
June 1, 2024 20:46 7m 25s master
June 1, 2024 20:46 7m 25s
ggml: implement quantized KV cache for FA (#7372)
Publish Docker image #56: Commit 5ca49cb pushed by katsu560
May 19, 2024 15:05 7m 18s master
May 19, 2024 15:05 7m 18s
Reset schedule earlier to allow overlap with ggml graph computation o…
Publish Docker image #55: Commit 928e0b7 pushed by katsu560
April 27, 2024 02:09 9m 36s master
April 27, 2024 02:09 9m 36s
ci: bench: support sse and fix prompt processing time / server: add t…
Publish Docker image #54: Commit 75cd4c7 pushed by katsu560
April 6, 2024 13:00 6m 41s master
April 6, 2024 13:00 6m 41s
gitignore : gguf-split
Publish Docker image #53: Commit 9556217 pushed by katsu560
March 23, 2024 21:11 6m 51s master
March 23, 2024 21:11 6m 51s
readme : add API changes section
Publish Docker image #52: Commit 231ae28 pushed by katsu560
March 3, 2024 11:54 9m 14s master
March 3, 2024 11:54 9m 14s
mpt : do not duplicate token_embd.weight on disk (#5670)
Publish Docker image #51: Commit 15499eb pushed by katsu560
February 23, 2024 14:05 6m 28s master
February 23, 2024 14:05 6m 28s
cmake : fix VULKAN and ROCm builds (#5525)
Publish Docker image #50: Commit 5bf2b94 pushed by katsu560
February 17, 2024 16:25 8m 21s master
February 17, 2024 16:25 8m 21s
vulkan: Set limit for task concurrency (#5427)
Publish Docker image #49: Commit 4b7b38b pushed by katsu560
February 10, 2024 00:31 16m 38s master
February 10, 2024 00:31 16m 38s
readme : add phi, orion 14b, internlm2, and yi-VL to readme (#5362)
Publish Docker image #48: Commit 2e9c0bd pushed by katsu560
February 6, 2024 14:15 6m 36s master
February 6, 2024 14:15 6m 36s
Remove unused data and add fixes (#5154)
Publish Docker image #47: Commit 35a2ee9 pushed by katsu560
January 27, 2024 14:53 12m 52s master
January 27, 2024 14:53 12m 52s
llama.swiftui : use correct pointer for llama_token_eos (#4797)
Publish Docker image #46: Commit c75ca5d pushed by katsu560
January 7, 2024 03:17 7m 42s master
January 7, 2024 03:17 7m 42s
flake.lock: update
Publish Docker image #45: Commit edd1ab7 pushed by katsu560
January 1, 2024 03:31 6m 1s master
January 1, 2024 03:31 6m 1s
fallback to CPU buffer if host buffer alloc fails (#4610)
Publish Docker image #44: Commit 708e179 pushed by katsu560
December 24, 2023 01:11 6m 45s master
December 24, 2023 01:11 6m 45s
llama : sanity checks for access to logits (#4274)
Publish Docker image #43: Commit 8a5be3b pushed by katsu560
December 16, 2023 13:29 4m 6s master
December 16, 2023 13:29 4m 6s
llama : avoid using "optional" keyword (#4283)
Publish Docker image #42: Commit 5a7d312 pushed by katsu560
December 2, 2023 06:53 3m 46s master
December 2, 2023 06:53 3m 46s
server : allow continue edit on completion mode (#3950)
Publish Docker image #41: Commit 4a4fd3e pushed by katsu560
November 11, 2023 03:33 3m 48s master
November 11, 2023 03:33 3m 48s