bhavyaaiplanet commited on
Commit
d0e253f
1 Parent(s): 16ae92f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -12
README.md CHANGED
@@ -30,15 +30,15 @@ effi 7b GPTQ is a quantized version of effi 7b whiich is a 7 billion parameter m
30
 
31
  ### Qunatization Configuration
32
 
33
- - bits: 4,
34
- - damp_percent: 0.1,
35
- - dataset: "wikitext2",
36
- - desc_act: false,
37
- - group_size: 128,
38
- - modules_in_block_to_quantize: null,
39
- - quant_method: "gptq",
40
- - sym: true,
41
- - true_sequential: true
42
 
43
  ### Example of usage
44
 
@@ -79,9 +79,9 @@ print(f"{tokenizer.batch_decode(outputs.detach().cpu().numpy(), skip_special_tok
79
  ```
80
 
81
  ### Framework versions
82
- - Transformers 4.37.2
83
- - optimum 1.16.2
84
- - auto-gptq 0.6.0
85
 
86
  ### Citation
87
 
 
30
 
31
  ### Qunatization Configuration
32
 
33
+ - **bits:** 4,
34
+ - **damp_percent** 0.1,
35
+ - **dataset:** "wikitext2",
36
+ - **desc_act:** false,
37
+ - **group_size:** 128,
38
+ - **modules_in_block_to_quantize:** null,
39
+ - **quant_method:** "gptq",
40
+ - **sym:** true,
41
+ - **true_sequential:** true
42
 
43
  ### Example of usage
44
 
 
79
  ```
80
 
81
  ### Framework versions
82
+ - **Transformers** 4.37.2
83
+ - **optimum** 1.16.2
84
+ - **auto-gptq** 0.6.0
85
 
86
  ### Citation
87