Skip to content
GitLab
Explore
Sign in
Overview
Active
Stale
All
This project is mirrored from
https://github.com/ggerganov/llama.cpp
. Pull mirroring updated
Sep 20, 2024
.
cuda-dmmv-dims
166e44b7
·
ggml-cuda : move row numbers to x grid dim in mmv kernels
·
Nov 03, 2023
fix-cuda-warnings
d1a1678b
·
Merge branch 'master' into fix-cuda-warnings
·
Nov 02, 2023
gguf-grace
f3069478
·
gguf : print error for GGUFv1 files
·
Nov 02, 2023
disable-cmake-native
c217a66b
·
disable LLAMA_NATIVE by default
·
Nov 02, 2023
remove-ggufv1
347d587a
·
gguf : remove special-case code for GGUFv1
·
Nov 02, 2023
fix-metal-after-yarn
396412c0
·
metal : fix build errors and kernel sig after #2268
·
Nov 02, 2023
metal-soft-max
46868a49
·
metal : multi-simd softmax
·
Nov 01, 2023
batched-krn
1354122c
·
fix warnings
·
Nov 01, 2023
llm-build-context
a8796f96
·
llm : cleanup + comments
·
Nov 01, 2023
llm-reuse-constants
7420bef8
·
wip wip wip
·
Nov 01, 2023
llama-refactor
afb39292
·
Merge branch 'master' into llama-refactor
·
Oct 31, 2023
try-fix-3869
22cc9bef
·
cuda : check if this fixes Pascal card regression
·
Oct 31, 2023
test-mmv
29fe5169
·
wip
·
Oct 31, 2023
deploy
dab42893
·
scripts : working curl pipe
·
Oct 31, 2023
llama-refactor-norm
7923b70c
·
llama : add llm_build_inp_embd helper
·
Oct 31, 2023
ggml-impl
4b3cb98d
·
ggml-impl : move extern "C" to start of file
·
Oct 30, 2023
lto
bc28aaa8
·
make : use -lfto=auto to avoid warnings and maintain perf
·
Oct 30, 2023
llama-refactor-ffn
3b778a4a
·
llama : add llm_build_ffn helper function
·
Oct 29, 2023
scratch
15267192
·
llama : refactor tensor offloading as callback
·
Oct 29, 2023
ggml-quants
8a86b95e
·
quantize : --pure option for disabling k-quant mixtures
·
Oct 28, 2023
Prev
1
…
11
12
13
14
15
16
17
18
19
…
26
Next