ymcki commited on
Commit
50f7347
1 Parent(s): cf95426

Upload 7 files

Browse files
config.json CHANGED
@@ -27,7 +27,7 @@
27
  "rope_theta": 10000.0,
28
  "sliding_window": 4096,
29
  "torch_dtype": "bfloat16",
30
- "transformers_version": "4.45.1",
31
  "use_cache": true,
32
  "vocab_size": 256000
33
  }
 
27
  "rope_theta": 10000.0,
28
  "sliding_window": 4096,
29
  "torch_dtype": "bfloat16",
30
+ "transformers_version": "4.45.2",
31
  "use_cache": true,
32
  "vocab_size": 256000
33
  }
generation_config.json CHANGED
@@ -4,5 +4,5 @@
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 1,
6
  "pad_token_id": 0,
7
- "transformers_version": "4.45.1"
8
  }
 
4
  "cache_implementation": "hybrid",
5
  "eos_token_id": 1,
6
  "pad_token_id": 0,
7
+ "transformers_version": "4.45.2"
8
  }
model.safetensors.index.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
  "metadata": {
3
- "total_size": 6408331776
4
  },
5
  "weight_map": {
6
- "lm_head.weight": "model-00002-of-00002.safetensors",
7
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
9
  "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5228683776
4
  },
5
  "weight_map": {
 
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
7
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
8
  "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",