BlackSamorez commited on
Commit
4ae602a
β€’
1 Parent(s): 891bf06

new dispatch again + correct dtype

Browse files
config.json CHANGED
@@ -1,20 +1,10 @@
1
  {
2
- "_name_or_path": "/extra_disk_1/dkuznedelev/aqlm/Llama-2-70b-AQLM-2Bit-1x16-hf",
3
- "aqlm": {
4
- "in_group_size": 8,
5
- "nbits_per_codebook": 16,
6
- "num_codebooks": 1,
7
- "out_group_size": 1
8
- },
9
  "architectures": [
10
  "LlamaForCausalLM"
11
  ],
12
  "attention_bias": false,
13
  "attention_dropout": 0.0,
14
- "auto_map": {
15
- "AutoConfig": "configuration_llama_aqlm.LlamaConfig",
16
- "AutoModelForCausalLM": "modeling_llama_aqlm.LlamaForCausalLM"
17
- },
18
  "bos_token_id": 1,
19
  "eos_token_id": 2,
20
  "hidden_act": "silu",
@@ -22,17 +12,28 @@
22
  "initializer_range": 0.02,
23
  "intermediate_size": 28672,
24
  "max_position_embeddings": 4096,
25
- "model_type": "llama_aqlm",
26
  "num_attention_heads": 64,
27
  "num_hidden_layers": 80,
28
  "num_key_value_heads": 8,
29
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
30
  "rms_norm_eps": 1e-05,
31
  "rope_scaling": null,
32
  "rope_theta": 10000.0,
33
  "tie_word_embeddings": false,
34
- "torch_dtype": "float32",
35
- "transformers_version": "4.37.2",
36
  "use_cache": true,
37
  "vocab_size": 32000
38
  }
 
1
  {
2
+ "_name_or_path": "BlackSamorez/Llama-2-70b-AQLM-2Bit-1x16-hf",
 
 
 
 
 
 
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
 
 
 
8
  "bos_token_id": 1,
9
  "eos_token_id": 2,
10
  "hidden_act": "silu",
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 28672,
14
  "max_position_embeddings": 4096,
15
+ "model_type": "llama",
16
  "num_attention_heads": 64,
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "pretraining_tp": 1,
20
+ "quantization_config": {
21
+ "in_group_size": 8,
22
+ "linear_weights_not_to_quantize": [
23
+ "model.embed_tokens.weight",
24
+ "lm_head.weight"
25
+ ],
26
+ "nbits_per_codebook": 16,
27
+ "num_codebooks": 1,
28
+ "out_group_size": 1,
29
+ "quant_method": "aqlm"
30
+ },
31
  "rms_norm_eps": 1e-05,
32
  "rope_scaling": null,
33
  "rope_theta": 10000.0,
34
  "tie_word_embeddings": false,
35
+ "torch_dtype": "float16",
36
+ "transformers_version": "4.38.0",
37
  "use_cache": true,
38
  "vocab_size": 32000
39
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.37.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.38.0"
6
  }
model-00001-of-00005.safetensors β†’ model-00001-of-00004.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9dc9c6d4d68cb998a9602f81f240e0fb1ffc333aeabed3746f4af44d154c51e
3
- size 4989778312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60e85d108d68e360d0342453e2106d43d67bbc883923b138f9fe7c2f792f7f76
3
+ size 4996376224
model-00002-of-00005.safetensors β†’ model-00002-of-00004.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ca83217933c3d4aa734853be1ed897bd4f375e48fbb2dee351c15e0b20e7978
3
- size 4991647616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63f1e45dca214aa63179040ac183b2a90468a30f1a7e4b6aedd463369bee5688
3
+ size 4991616704
model-00003-of-00005.safetensors β†’ model-00003-of-00004.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c1f1e4dd22423b61d6a3550cc78af488f35a1926aa218a48ddb07879db2da29
3
- size 4976943808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e66eefba897aedfcecec42a2d89286e8e17a364ee6734649471ebc85c5ff51b
3
+ size 4973763520
model-00004-of-00005.safetensors β†’ model-00004-of-00004.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:449da0466ca4bc5eec0a0a557319a27794c8743002a6055dccf3e5e5c3b6498c
3
- size 4409730992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d808e60f33b9aac1c9a75e6def7f9762061b6487f9484a3f6ca3af2940f37087
3
+ size 3803068312
model-00005-of-00005.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:34576756203b32b277f651fc65fa3f8f7171e7bb5796ec52bbeb4e5bb4a645af
3
- size 1048576128
 
 
 
 
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff