diff options
author | Georgi Gerganov <ggerganov@gmail.com> | 2023-03-12 08:41:54 +0200 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-03-12 08:41:54 +0200 |
commit | 7d86e25bf648eb369a3a8388bf239b6b19f7a789 (patch) | |
tree | d3e0b85ab3f4b4d95ac23f682709f3200022bcf9 /README.md | |
parent | a93120236f99e13d77e4b278e47ffcaad4a899e4 (diff) |
README: add "Supported platforms" + update hot topics
Diffstat (limited to 'README.md')
-rw-r--r-- | README.md | 9 |
1 files changed, 8 insertions, 1 deletions
@@ -5,10 +5,11 @@ Inference of [Facebook's LLaMA](https://github.com/facebookresearch/llama) model **Hot topics** - Running on Windows: https://github.com/ggerganov/llama.cpp/issues/22 +- Fix Tokenizer / Unicode support: https://github.com/ggerganov/llama.cpp/issues/11 ## Description -The main goal is to run the model using 4-bit quantization on a MacBook. +The main goal is to run the model using 4-bit quantization on a MacBook - Plain C/C++ implementation without dependencies - Apple silicon first-class citizen - optimized via Arm Neon and Accelerate framework @@ -22,6 +23,12 @@ Please do not make conclusions about the models based on the results from this i For all I know, it can be completely wrong. This project is for educational purposes and is not going to be maintained properly. New features will probably be added mostly through community contributions, if any. +Supported platformst: + +- [X] Mac OS +- [X] Linux +- [ ] Windows (soon) + --- Here is a typical run using LLaMA-7B: |