Skip to content
GitLab
Explore
Sign in
Overview
Active
Stale
All
This project is mirrored from
https://github.com/ggerganov/llama.cpp
. Pull mirroring updated
Sep 19, 2024
.
gg/flash-attn-rebase
3a468e6f
·
llama : fix type of KQ_mask and KQ_pos
·
Mar 22, 2024
gg/flash-attn
09532120
·
ggml : fix CPU soft_max
·
Mar 22, 2024
hp/server/logs-flush
2187f34b
·
server: flush stdout after logging in both text and json layout
·
Mar 23, 2024
ceb/fix-win-unicode-fpaths
d05c13b3
·
llama : fix BPE LF token on MSVC
·
Mar 23, 2024
update_flake_lock_action
diverged from upstream
038a851d
·
flake.lock: Update
·
Mar 24, 2024
sycl-offload-op
5f8a87d7
·
remove sycl part from common backend
·
Mar 24, 2024
sl/cuda-refactor-files
0f304d9b
·
cuda : refactor into multiple files
·
Mar 24, 2024
revert-6572-master
92093acf
·
Revert "minor layout improvements (#6572)"
·
Apr 15, 2024
revert-6951-master
0b7c4e82
·
Revert "move ndk code to a new library (#6951)"
·
May 14, 2024
revert-7682-master
315c3afe
·
Revert "[SYCL] Update rpc-server.cpp to include SYCL backend (#7682)"
·
Jun 06, 2024
Prev
1
…
22
23
24
25
26
Next