lex-hue commited on
Commit
2d38420
1 Parent(s): f80f304

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -4,20 +4,31 @@
4
  "MistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
 
 
 
 
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 14336,
13
- "max_position_embeddings": 8192,
 
14
  "model_type": "mistral",
15
  "num_attention_heads": 32,
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 8,
18
  "rms_norm_eps": 1e-05,
 
 
 
 
 
 
19
  "rope_theta": 10000.0,
20
- "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
  "transformers_version": "4.39.3",
 
4
  "MistralForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_mistral.MistralConfig",
9
+ "AutoModelForCausalLM": "modeling_mistral_yarn.MistralForCausalLM"
10
+ },
11
  "bos_token_id": 1,
12
  "eos_token_id": 2,
13
  "hidden_act": "silu",
14
  "hidden_size": 4096,
15
  "initializer_range": 0.02,
16
  "intermediate_size": 14336,
17
+ "max_position_embeddings": 32768,
18
+ "max_sequence_length": 131072,
19
  "model_type": "mistral",
20
  "num_attention_heads": 32,
21
  "num_hidden_layers": 32,
22
  "num_key_value_heads": 8,
23
  "rms_norm_eps": 1e-05,
24
+ "rope_scaling": {
25
+ "factor": 16.0,
26
+ "finetuned": true,
27
+ "original_max_position_embeddings": 8192,
28
+ "type": "yarn"
29
+ },
30
  "rope_theta": 10000.0,
31
+ "sliding_window": 131072,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "float16",
34
  "transformers_version": "4.39.3",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc48dfefbb790a69b82b078085e99150eb47cb69784b70ef71fc1029252af1a9
3
  size 9825524256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0aa3052a42f9e13b77eaed24bbadfc72414cdb43b9aecbf5e30bc273483c2ff
3
  size 9825524256
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86caa904f43f806f27fec47fb3ecec05135b957a8fc554e42699b29657a87a86
3
  size 4657973496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fda89f364a7e7eee50f31b8efb16561ad4d05c863e780ed806f5df3e6a2fb566
3
  size 4657973496
special_tokens_map.json CHANGED
@@ -1,4 +1,9 @@
1
  {
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<unk>",
4
+ "<s>",
5
+ "</s>"
6
+ ],
7
  "bos_token": {
8
  "content": "<s>",
9
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -27,7 +27,11 @@
27
  "special": true
28
  }
29
  },
30
- "additional_special_tokens": [],
 
 
 
 
31
  "bos_token": "<s>",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
@@ -38,5 +42,5 @@
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
40
  "unk_token": "<unk>",
41
- "use_default_system_prompt": false
42
  }
 
27
  "special": true
28
  }
29
  },
30
+ "additional_special_tokens": [
31
+ "<unk>",
32
+ "<s>",
33
+ "</s>"
34
+ ],
35
  "bos_token": "<s>",
36
  "clean_up_tokenization_spaces": false,
37
  "eos_token": "</s>",
 
42
  "spaces_between_special_tokens": false,
43
  "tokenizer_class": "LlamaTokenizer",
44
  "unk_token": "<unk>",
45
+ "use_default_system_prompt": true
46
  }