froggeric commited on
Commit
e8406a0
1 Parent(s): 0596a98

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md CHANGED
@@ -4,6 +4,25 @@ license_name: deepseek
4
  license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
5
  ---
6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  # This model is now live (We'll always be serving the newest model on our web app)!
9
  Access at: https://www.whiterabbitneo.com/
 
4
  license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
5
  ---
6
 
7
+ Q8_0 GGUF quantization of [WhiteRabbitNeo/WhiteRabbitNeo-33B-v1.5](https://huggingface.co/WhiteRabbitNeo/WhiteRabbitNeo-33B-v1.5)
8
+ converted and quantized with llama.cpp
9
+
10
+ Please note there is a **bug** in the `convert.py` script from recent versions of llama.cpp, that **affects models
11
+ with additional vocabulary** (like this one). This results in model that can be converted, but have some garbage
12
+ and cannot be used for inference. Until this is fixed, you must use an older version of `convert.py`
13
+ (only this script, the rest of llama.cpp can stay with the latest version), and use the following parameters:
14
+ `--pad-vocab --vocab-type bpe`. For example. with a locally downloaded version of the original model repo:
15
+
16
+ ```
17
+ python llama.cpp/convert-aa23412.py WhiteRabbitNeo-33B-v1.5 --pad-vocab --vocab-type bpe
18
+ ```
19
+
20
+ The **last working version** of `convert.py` is [aa23412](https://github.com/ggerganov/llama.cpp/tree/aa2341298924ac89778252015efcb792f2df1e20)
21
+
22
+ ---
23
+
24
+ # ORGINAL MODEL CARD
25
+
26
 
27
  # This model is now live (We'll always be serving the newest model on our web app)!
28
  Access at: https://www.whiterabbitneo.com/