aifeifei798 commited on
Commit
55b7e6c
1 Parent(s): 24ea3d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -16,7 +16,7 @@ tags:
16
 
17
  # fast quantizations
18
  - The difference with normal quantizations is that I quantize the output and embed tensors to f16.and the other tensors to 15_k,q6_k or q8_0.This creates models that are little or not degraded at all and have a smaller size.They run at about 3-6 t/sec on CPU only using llama.cpp And obviously faster on computers with potent GPUs
19
- - ZeroWw/llama3-8B-DarkIdol-2.1-Uncensored-32K-GGUF
20
  - More models here: https://huggingface.co/RobertSinclair
21
 
22
  # Model Description:
 
16
 
17
  # fast quantizations
18
  - The difference with normal quantizations is that I quantize the output and embed tensors to f16.and the other tensors to 15_k,q6_k or q8_0.This creates models that are little or not degraded at all and have a smaller size.They run at about 3-6 t/sec on CPU only using llama.cpp And obviously faster on computers with potent GPUs
19
+ - https://huggingface.co/ZeroWw/llama3-8B-DarkIdol-2.1-Uncensored-32K-GGUF
20
  - More models here: https://huggingface.co/RobertSinclair
21
 
22
  # Model Description: