Skip to content

Commit e9305da

Browse files
committed
Add the new data types across files
1 parent a386408 commit e9305da

File tree

3 files changed

+6
-2
lines changed

3 files changed

+6
-2
lines changed

examples/quantize/quantize.cpp

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,7 @@ struct quant_option {
1717

1818
static const std::vector<struct quant_option> QUANT_OPTIONS = {
1919
{ "Q4_0", LLAMA_FTYPE_MOSTLY_Q4_0, " 4.34G, +0.4685 ppl @ Llama-3-8B", },
20+
{ "Q4_0_B16", LLAMA_FTYPE_MOSTLY_Q4_0_B16, " 3.56G, 5.9624 +/- 0.03348 ppl @ LLaMA-v2-7B", },
2021
{ "Q4_1", LLAMA_FTYPE_MOSTLY_Q4_1, " 4.78G, +0.4511 ppl @ Llama-3-8B", },
2122
{ "Q5_0", LLAMA_FTYPE_MOSTLY_Q5_0, " 5.21G, +0.1316 ppl @ Llama-3-8B", },
2223
{ "Q5_1", LLAMA_FTYPE_MOSTLY_Q5_1, " 5.65G, +0.1062 ppl @ Llama-3-8B", },
@@ -46,6 +47,7 @@ static const std::vector<struct quant_option> QUANT_OPTIONS = {
4647
{ "Q5_K_M", LLAMA_FTYPE_MOSTLY_Q5_K_M, " 5.33G, +0.0569 ppl @ Llama-3-8B", },
4748
{ "Q6_K", LLAMA_FTYPE_MOSTLY_Q6_K, " 6.14G, +0.0217 ppl @ Llama-3-8B", },
4849
{ "Q8_0", LLAMA_FTYPE_MOSTLY_Q8_0, " 7.96G, +0.0026 ppl @ Llama-3-8B", },
50+
{ "Q8_0_B16", LLAMA_FTYPE_MOSTLY_Q8_0_B16, " 6.70G, 5.8011 +/- 0.03239 ppl @ LLaMA-v1-7B", },
4951
{ "Q4_0_4_4", LLAMA_FTYPE_MOSTLY_Q4_0_4_4, " 4.34G, +0.4685 ppl @ Llama-3-8B", },
5052
{ "Q4_0_4_8", LLAMA_FTYPE_MOSTLY_Q4_0_4_8, " 4.34G, +0.4685 ppl @ Llama-3-8B", },
5153
{ "Q4_0_8_8", LLAMA_FTYPE_MOSTLY_Q4_0_8_8, " 4.34G, +0.4685 ppl @ Llama-3-8B", },

gguf-py/gguf/constants.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1145,8 +1145,8 @@ class GGMLQuantizationType(IntEnum):
11451145
F64 = 28
11461146
IQ1_M = 29
11471147
BF16 = 30
1148-
Q4_0_B16 = 31
1149-
Q8_0_B16 = 32
1148+
Q4_0_B16 = 34
1149+
Q8_0_B16 = 35
11501150

11511151

11521152
# TODO: add GGMLFileType from ggml_ftype in ggml.h

src/llama.cpp

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4452,10 +4452,12 @@ static std::string llama_model_ftype_name(llama_ftype ftype) {
44524452
case LLAMA_FTYPE_MOSTLY_F16: return "F16";
44534453
case LLAMA_FTYPE_MOSTLY_BF16: return "BF16";
44544454
case LLAMA_FTYPE_MOSTLY_Q4_0: return "Q4_0";
4455+
case LLAMA_FTYPE_MOSTLY_Q4_0_B16: return "Q4_0_B16";
44554456
case LLAMA_FTYPE_MOSTLY_Q4_1: return "Q4_1";
44564457
case LLAMA_FTYPE_MOSTLY_Q5_0: return "Q5_0";
44574458
case LLAMA_FTYPE_MOSTLY_Q5_1: return "Q5_1";
44584459
case LLAMA_FTYPE_MOSTLY_Q8_0: return "Q8_0";
4460+
case LLAMA_FTYPE_MOSTLY_Q8_0_B16: return "Q4_0_B16";
44594461
case LLAMA_FTYPE_MOSTLY_Q2_K: return "Q2_K - Medium";
44604462
case LLAMA_FTYPE_MOSTLY_Q2_K_S: return "Q2_K - Small";
44614463
case LLAMA_FTYPE_MOSTLY_Q3_K_S: return "Q3_K - Small";

0 commit comments

Comments
 (0)