Skip to content
GitLab
Explore
Sign in
Overview
Active
Stale
All
This project is mirrored from
https://github.com/ggerganov/llama.cpp
. Pull mirroring updated
Sep 20, 2024
.
gg/hf-auto-dl
120a1a55
·
llama : auto download HF models if URL provided
·
Jan 02, 2024
gg/fix-mingw-4707
c92418dc
·
ggml : include stdlib.h before intrin.h
·
Jan 02, 2024
gg/avoid-mutex
b5af7ad8
·
llama : refactor quantization to avoid <mutex> header
·
Jan 02, 2024
cuda-cublas-opts
4cc78d38
·
ggml : force F32 precision for ggml_mul_mat
·
Jan 02, 2024
gg/metal-opt-mul-mat-id
9f51f3e6
·
metal : opt mul_mm_id
·
Jan 02, 2024
ik/iq2_2.06bpw
1e6b8e1f
·
iq2_xxs: add to llama ftype enum
·
Jan 04, 2024
gg/no-yield-on-blas
4a0e7222
·
ggml : simplify do_yield logic
·
Jan 04, 2024
gg/base-translate
26fbb10f
·
examples : add few-shot translation example
·
Jan 05, 2024
gg/remove-gqa-check-4657
7cfde781
·
llama : remove redundant GQA check
·
Jan 06, 2024
gg/fix-server-n-decoded-4790
58de6736
·
server : fix n_predict check
·
Jan 06, 2024
sl/backend-sched-fix-null-deref
44c93c67
·
ggml : also check ggml_backend_is_cpu
·
Jan 07, 2024
gg/self-extend
f33a8106
·
passkey : add comment
·
Jan 07, 2024
gg/self-extend-part-2
ea129218
·
main : add Self-Extend support
·
Jan 07, 2024
passkey
d57cb9c2
·
passkey : add readme
·
Jan 08, 2024
gg/exclude-resources-spm
4aa73e37
·
swift : exclude ggml-metal.metal from the package
·
Jan 08, 2024
gg/fix-vld1q_s8_x4-4872
7216af5c
·
ggml : fix 32-bit ARM compat (cont)
·
Jan 09, 2024
gg/server-infill-empty-prompt-4027
24096933
·
server : try to fix infill when prompt is empty
·
Jan 09, 2024
gg/metal-ci-fixes
ef8ba127
·
metal : improve dequantize precision to match CPU
·
Jan 09, 2024
sl/backend-sched-page-align
2063b868
·
metal : page align the data ptr
·
Jan 10, 2024
ik/imatrix
f0b71d5d
·
Cleanup
·
Jan 10, 2024
Prev
1
…
12
13
14
15
16
17
18
19
20
…
26
Next