index
:
llama.cpp.git
master
llama.cpp
user
about
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
examples
/
quantize
Age
Commit message (
Expand
)
Author
2023-05-05
quantize: make output filename optional, default to ggml-model-<ftype>.bin (#...
slaren
2023-05-01
Add git-based build information for better issue tracking (#1232)
DannyDaemonic
2023-04-28
Remove Q4_3 which is no better than Q5 (#1218)
Stephan Walter
2023-04-26
ggml : add Q5_0 and Q5_1 quantization (#1187)
Georgi Gerganov
2023-04-26
quantize : use `map` to assign quantization type from `string` (#1191)
Pavol Rusnak
2023-04-25
ggml : add Q8_0 quantization format (rename the old one to Q8_1) (ARM NEON) (...
Georgi Gerganov
2023-04-20
llama : multi-threaded quantization (#1075)
Kawrakow
2023-04-20
ggml : add Q4_3 quantization (#1082)
Georgi Gerganov
2023-04-18
ggml : add new Q4_2 quantization (ARM only) (#1046)
Georgi Gerganov
2023-04-11
Add enum llama_ftype, sync ggml_type to model files (#709)
Stephan Walter
2023-03-30
Fix ggml_init_params in quantize
Slaren
2023-03-28
llama : fix linkage with mingw (#551)
anzz1
2023-03-28
all : be more strict about converting float to double (#458)
Stephan Walter
2023-03-28
ggml : introduce structs for the q4 data blocks (#356)
Stephan Walter
2023-03-27
Fix missing ggml link in cmake for examples/* on w64-mingw32 (#542)
Marco Matthies
2023-03-25
Overhaul the examples structure
Georgi Gerganov