Update README.md
Browse files
README.md
CHANGED
@@ -30,12 +30,11 @@ license: cc-by-nc-4.0
|
|
30 |
- Requantized for recent bpe pre-tokenizer fixes https://github.com/ggerganov/llama.cpp/pull/6920
|
31 |
- quants done with an importance matrix for improved quantization loss
|
32 |
- 0, K & IQ quants in basically all variants from Q8 down to IQ1_S
|
33 |
-
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [
|
34 |
- Imatrix generated with [this](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) dataset.
|
35 |
```
|
36 |
./imatrix -c 512 -m $model_name-f16.gguf -f $llama_cpp_path/groups_merged.txt -o $out_path/imat-f16-gmerged.dat
|
37 |
```
|
38 |
-
Quantized with llama.cpp commit fabf30b4c4fca32e116009527180c252919ca922 (master as of 2024-05-20)
|
39 |
|
40 |
# Original Model Card:
|
41 |
|
|
|
30 |
- Requantized for recent bpe pre-tokenizer fixes https://github.com/ggerganov/llama.cpp/pull/6920
|
31 |
- quants done with an importance matrix for improved quantization loss
|
32 |
- 0, K & IQ quants in basically all variants from Q8 down to IQ1_S
|
33 |
+
- Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [fabf30b4c4fca32e116009527180c252919ca922](https://github.com/ggerganov/llama.cpp/commit/fabf30b4c4fca32e116009527180c252919ca922) (master as of 2024-05-20)
|
34 |
- Imatrix generated with [this](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384) dataset.
|
35 |
```
|
36 |
./imatrix -c 512 -m $model_name-f16.gguf -f $llama_cpp_path/groups_merged.txt -o $out_path/imat-f16-gmerged.dat
|
37 |
```
|
|
|
38 |
|
39 |
# Original Model Card:
|
40 |
|