Skip to content
GitLab
Explore
Sign in
Overview
Active
Stale
All
This project is mirrored from
https://github.com/ggerganov/llama.cpp
. Pull mirroring updated
Sep 19, 2024
.
gguf-publish-ci
488e0320
·
Merge branch 'master' into gguf-publish-ci
·
Aug 30, 2023
fix-docker-tools-sh
72177120
·
remove `exec`
·
Aug 29, 2023
ik/issue_2858
c6b8bdbc
·
Tell users attmepting to run perplexity with too few tokens to use more
·
Aug 29, 2023
llama2-readme
4a4051b8
·
remove outdated references to -eps and -gqa from README
·
Aug 29, 2023
view-src
9fca82be
·
formatting
·
Aug 29, 2023
ik/faster_bpe_tokenizer
849a31f1
·
Remove comment that no longer applies
·
Aug 29, 2023
fix-gguf-str
74999e08
·
gguf : fix strings to not be null-terminated
·
Aug 27, 2023
add-abort-callback
f2770b8c
·
Add abort callback
·
Aug 27, 2023
ik/speedup_tokenization
86e35115
·
Fixit: it was missing the piece after the last found occurence
·
Aug 27, 2023
llama-bench-utf8
e000ff7b
·
llama-bench : set locale to utf8
·
Aug 27, 2023
fix-falcon
cc924c57
·
cuda : add assert to guard from non-cont ropes
·
Aug 27, 2023
gguf-64bit
33a5517d
·
llama.cpp : print gguf version
·
Aug 26, 2023
fix-tokenizer
724fa67d
·
perplexity.cpp : better way to deal wirh spm prepending space
·
Aug 26, 2023
ik/fix_cuda_qkk64
f547c585
·
Make ggml-cuda.cu build with QK_K = 64
·
Aug 26, 2023
ik/refine_70B
3979af1e
·
PR comment
·
Aug 26, 2023
fix-code-llama-quantum-mixtures
eff86d4f
·
k-quants : remove unnecessary tensor shape restrictions
·
Aug 26, 2023
ik/fix_hellaswag
d34472c1
·
Fix HellaSwag
·
Aug 26, 2023
ik/faster_ppl
ce45974a
·
Faster perplexity computation
·
Aug 25, 2023
llama-bench-model-size
bc0dc16c
·
more compact markdown output
·
Aug 25, 2023
ci-gguf
e38f7847
·
ci : pip install gguf in editable mode
·
Aug 25, 2023
Prev
1
…
16
17
18
19
20
21
22
23
24
…
26
Next