diff options
author | Mack Straight <eiz@users.noreply.github.com> | 2023-03-20 03:17:23 -0700 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-03-20 03:17:23 -0700 |
commit | 074bea2eb1f1349a0118239c4152914aecaa1be4 (patch) | |
tree | 41ce911ac28d858cabfeff650b10521b30838656 /README.md | |
parent | 5cb63e2493c49bc2c3b9b355696e8dc26cdd0380 (diff) |
sentencepiece bpe compatible tokenizer (#252)
* potential out of bounds read
* fix quantize
* style
* Update convert-pth-to-ggml.py
* mild cleanup
* don't need the space-prefixing here rn since main.cpp already does it
* new file magic + version header field
* readme notice
* missing newlines
Co-authored-by: slaren <2141330+slaren@users.noreply.github.com>
Diffstat (limited to 'README.md')
-rw-r--r-- | README.md | 3 |
1 files changed, 3 insertions, 0 deletions
@@ -11,6 +11,9 @@ Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++ - Cache input prompts for faster initialization: https://github.com/ggerganov/llama.cpp/issues/64 - Create a `llama.cpp` logo: https://github.com/ggerganov/llama.cpp/issues/105 +**TEMPORARY NOTICE:** +If you're updating to the latest master, you will need to regenerate your model files as the format has changed. + ## Description The main goal is to run the model using 4-bit quantization on a MacBook |