aboutsummaryrefslogtreecommitdiff
path: root/examples/quantize/quantize.cpp
diff options
context:
space:
mode:
authorGeorgi Gerganov <ggerganov@gmail.com>2023-07-18 11:50:49 +0300
committerGeorgi Gerganov <ggerganov@gmail.com>2023-07-18 11:50:49 +0300
commit6cbf9dfb32f0e23ed3afd02d30ab066ed53e2c4d (patch)
tree6b6509004938369098887dee6d6499005bed93cb /examples/quantize/quantize.cpp
parent7568d1a2b206331412106ea66da3f871025e0c3c (diff)
llama : shorten quantization descriptions
Diffstat (limited to 'examples/quantize/quantize.cpp')
-rw-r--r--examples/quantize/quantize.cpp114
1 files changed, 19 insertions, 95 deletions
diff --git a/examples/quantize/quantize.cpp b/examples/quantize/quantize.cpp
index 797d2f0..744f549 100644
--- a/examples/quantize/quantize.cpp
+++ b/examples/quantize/quantize.cpp
@@ -14,103 +14,27 @@ struct quant_option {
};
static const std::vector<struct quant_option> QUANT_OPTIONS = {
- {
- "Q4_0",
- LLAMA_FTYPE_MOSTLY_Q4_0,
- " 3.50G, +0.2499 ppl @ 7B - small, very high quality loss - legacy, prefer using Q3_K_M",
- },
- {
- "Q4_1",
- LLAMA_FTYPE_MOSTLY_Q4_1,
- " 3.90G, +0.1846 ppl @ 7B - small, substantial quality loss - legacy, prefer using Q3_K_L",
- },
- {
- "Q5_0",
- LLAMA_FTYPE_MOSTLY_Q5_0,
- " 4.30G, +0.0796 ppl @ 7B - medium, balanced quality - legacy, prefer using Q4_K_M",
- },
- {
- "Q5_1",
- LLAMA_FTYPE_MOSTLY_Q5_1,
- " 4.70G, +0.0415 ppl @ 7B - medium, low quality loss - legacy, prefer using Q5_K_M",
- },
+ { "Q4_0", LLAMA_FTYPE_MOSTLY_Q4_0, " 3.50G, +0.2499 ppl @ 7B", },
+ { "Q4_1", LLAMA_FTYPE_MOSTLY_Q4_1, " 3.90G, +0.1846 ppl @ 7B", },
+ { "Q5_0", LLAMA_FTYPE_MOSTLY_Q5_0, " 4.30G, +0.0796 ppl @ 7B", },
+ { "Q5_1", LLAMA_FTYPE_MOSTLY_Q5_1, " 4.70G, +0.0415 ppl @ 7B", },
#ifdef GGML_USE_K_QUANTS
- {
- "Q2_K",
- LLAMA_FTYPE_MOSTLY_Q2_K,
- " 2.67G, +0.8698 ppl @ 7B - smallest, extreme quality loss - not recommended",
- },
- {
- "Q3_K",
- LLAMA_FTYPE_MOSTLY_Q3_K_M,
- "alias for Q3_K_M"
- },
- {
- "Q3_K_S",
- LLAMA_FTYPE_MOSTLY_Q3_K_S,
- " 2.75G, +0.5505 ppl @ 7B - very small, very high quality loss",
- },
- {
- "Q3_K_M",
- LLAMA_FTYPE_MOSTLY_Q3_K_M,
- " 3.06G, +0.2437 ppl @ 7B - very small, very high quality loss",
- },
- {
- "Q3_K_L",
- LLAMA_FTYPE_MOSTLY_Q3_K_L,
- " 3.35G, +0.1803 ppl @ 7B - small, substantial quality loss",
- },
- {
- "Q4_K",
- LLAMA_FTYPE_MOSTLY_Q4_K_M,
- "alias for Q4_K_M",
- },
- {
- "Q4_K_S",
- LLAMA_FTYPE_MOSTLY_Q4_K_S,
- " 3.56G, +0.1149 ppl @ 7B - small, significant quality loss",
- },
- {
- "Q4_K_M",
- LLAMA_FTYPE_MOSTLY_Q4_K_M,
- " 3.80G, +0.0535 ppl @ 7B - medium, balanced quality - *recommended*",
- },
- {
- "Q5_K",
- LLAMA_FTYPE_MOSTLY_Q5_K_M,
- "alias for Q5_K_M",
- },
- {
- "Q5_K_S",
- LLAMA_FTYPE_MOSTLY_Q5_K_S,
- " 4.33G, +0.0353 ppl @ 7B - large, low quality loss - *recommended*",
- },
- {
- "Q5_K_M",
- LLAMA_FTYPE_MOSTLY_Q5_K_M,
- " 4.45G, +0.0142 ppl @ 7B - large, very low quality loss - *recommended*",
- },
- {
- "Q6_K",
- LLAMA_FTYPE_MOSTLY_Q6_K,
- " 5.15G, +0.0044 ppl @ 7B - very large, extremely low quality loss",
- },
+ { "Q2_K", LLAMA_FTYPE_MOSTLY_Q2_K, " 2.67G, +0.8698 ppl @ 7B", },
+ { "Q3_K", LLAMA_FTYPE_MOSTLY_Q3_K_M, "alias for Q3_K_M" },
+ { "Q3_K_S", LLAMA_FTYPE_MOSTLY_Q3_K_S, " 2.75G, +0.5505 ppl @ 7B", },
+ { "Q3_K_M", LLAMA_FTYPE_MOSTLY_Q3_K_M, " 3.06G, +0.2437 ppl @ 7B", },
+ { "Q3_K_L", LLAMA_FTYPE_MOSTLY_Q3_K_L, " 3.35G, +0.1803 ppl @ 7B", },
+ { "Q4_K", LLAMA_FTYPE_MOSTLY_Q4_K_M, "alias for Q4_K_M", },
+ { "Q4_K_S", LLAMA_FTYPE_MOSTLY_Q4_K_S, " 3.56G, +0.1149 ppl @ 7B", },
+ { "Q4_K_M", LLAMA_FTYPE_MOSTLY_Q4_K_M, " 3.80G, +0.0535 ppl @ 7B", },
+ { "Q5_K", LLAMA_FTYPE_MOSTLY_Q5_K_M, "alias for Q5_K_M", },
+ { "Q5_K_S", LLAMA_FTYPE_MOSTLY_Q5_K_S, " 4.33G, +0.0353 ppl @ 7B", },
+ { "Q5_K_M", LLAMA_FTYPE_MOSTLY_Q5_K_M, " 4.45G, +0.0142 ppl @ 7B", },
+ { "Q6_K", LLAMA_FTYPE_MOSTLY_Q6_K, " 5.15G, +0.0044 ppl @ 7B", },
#endif
- {
- "Q8_0",
- LLAMA_FTYPE_MOSTLY_Q8_0,
- " 6.70G, +0.0004 ppl @ 7B - very large, extremely low quality loss - not recommended",
- },
- {
- "F16",
- LLAMA_FTYPE_MOSTLY_F16,
- "13.00G @ 7B - extremely large, virtually no quality loss - not recommended",
- },
- {
- "F32",
- LLAMA_FTYPE_ALL_F32,
- "26.00G @ 7B - absolutely huge, lossless - not recommended",
- },
+ { "Q8_0", LLAMA_FTYPE_MOSTLY_Q8_0, " 6.70G, +0.0004 ppl @ 7B", },
+ { "F16", LLAMA_FTYPE_MOSTLY_F16, "13.00G @ 7B", },
+ { "F32", LLAMA_FTYPE_ALL_F32, "26.00G @ 7B", },
};