Skip to content

Actions: ggerganov/llama.cpp

Server

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
9,029 workflow runs
9,029 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

llama : functions -> methods
Server #9535: Pull request #11110 opened by ggerganov
January 6, 2025 14:14 Queued gg/llama-refactor-7
January 6, 2025 14:14 Queued
llama : remove unused headers
Server #9534: Pull request #11109 opened by ggerganov
January 6, 2025 13:56 Queued gg/llama-refactor-6
January 6, 2025 13:56 Queued
server : fix extra BOS in infill endpoint (#11106)
Server #9533: Commit e6e7c75 pushed by ggerganov
January 6, 2025 13:36 45m 17s master
January 6, 2025 13:36 45m 17s
rpc : code cleanup
Server #9532: Pull request #11107 opened by rgerganov
January 6, 2025 13:17 8m 4s rgerganov:rpc-cleanup
January 6, 2025 13:17 8m 4s
server : fix extra BOS in infill endpoing
Server #9531: Pull request #11106 synchronize by ggerganov
January 6, 2025 12:56 19m 1s gg/server-infill-fix-bos
January 6, 2025 12:56 19m 1s
server : fix extra BOS in infill endpoing
Server #9530: Pull request #11106 opened by ggerganov
January 6, 2025 12:47 9m 32s gg/server-infill-fix-bos
January 6, 2025 12:47 9m 32s
llama : remove check flash_attn with lora (#11104)
Server #9529: Commit 09186fa pushed by ngxson
January 6, 2025 12:41 16m 32s master
January 6, 2025 12:41 16m 32s
server : POC OAI-compat TTS using OuteTTS
Server #9528: Pull request #11070 synchronize by ggerganov
January 6, 2025 11:36 1h 11m 27s ngxson:xsn/server_tts
January 6, 2025 11:36 1h 11m 27s
llama : remove check flash_attn with lora
Server #9527: Pull request #11104 opened by ngxson
January 6, 2025 11:29 59m 4s ngxson:xsn/flash_attn_lora
January 6, 2025 11:29 59m 4s
llama : prevent system info string accumulation across calls (#11101)
Server #9526: Commit 96a1dc2 pushed by ggerganov
January 6, 2025 11:21 1h 1m 47s master
January 6, 2025 11:21 1h 1m 47s
feat(ci): add visionOS build workflow
Server #9525: Pull request #11103 opened by ggerganov
January 6, 2025 11:05 1h 5m 6s shards-lang/gio/visionos-ci
January 6, 2025 11:05 1h 5m 6s
llama : rename missed batch params/vars to ubatch (#10059)
Server #9523: Commit 6369f86 pushed by ggerganov
January 6, 2025 09:28 2h 15m 32s master
January 6, 2025 09:28 2h 15m 32s
llama : remove notion of CLS token
Server #9522: Pull request #11064 synchronize by ggerganov
January 6, 2025 08:58 2h 22m 3s gg/llama-refactor-5
January 6, 2025 08:58 2h 22m 3s
llama : update llama_model API names (#11063)
Server #9521: Commit 47182dd pushed by ggerganov
January 6, 2025 08:55 2h 4m 47s master
January 6, 2025 08:55 2h 4m 47s
tokenize : escape the prompt (#11058)
Server #9520: Commit 3e6e7a6 pushed by ggerganov
January 6, 2025 08:54 1h 53m 55s master
January 6, 2025 08:54 1h 53m 55s
tokenize : escape the prompt
Server #9519: Pull request #11058 synchronize by ggerganov
January 6, 2025 08:54 1h 42m 0s gg/tokenize-escape
January 6, 2025 08:54 1h 42m 0s
mmap : fix fileno macro clash (#11076)
Server #9518: Commit ae2f606 pushed by ggerganov
January 6, 2025 08:52 1h 19m 46s master
January 6, 2025 08:52 1h 19m 46s
llama : use LLAMA_TOKEN_NULL (#11062)
Server #9517: Commit 727368c pushed by ggerganov
January 6, 2025 08:52 1h 2m 20s master
January 6, 2025 08:52 1h 2m 20s
llama : use _impl suffix instead of _internal (#11060)
Server #9516: Commit 5047dd3 pushed by ggerganov
January 6, 2025 08:52 1h 0m 7s master
January 6, 2025 08:52 1h 0m 7s
Remove obsolete HIP workaround
Server #9515: Pull request #11080 synchronize by sARY77
January 6, 2025 04:36 Action required sARY77:Remove_obsolete_HIP_workaround
January 6, 2025 04:36 Action required
fix: Vulkan shader gen binary path when Cross-compiling
Server #9514: Pull request #11096 synchronize by ag2s20150909
January 6, 2025 01:56 5m 55s ag2s20150909:patch-2
January 6, 2025 01:56 5m 55s
fix: Vulkan shader gen binary path when Cross-compiling
Server #9513: Pull request #11096 synchronize by ag2s20150909
January 6, 2025 01:49 6m 25s ag2s20150909:patch-2
January 6, 2025 01:49 6m 25s
CUDA: add BF16 support (#11093)
Server #9511: Commit 46e3556 pushed by JohannesGaessler
January 6, 2025 01:33 6m 30s master
January 6, 2025 01:33 6m 30s