teknium commited on
Commit
66b73e7
1 Parent(s): f0de204

Upload 5 files

Browse files
config.json CHANGED
@@ -1,11 +1,12 @@
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 128000,
8
- "eos_token_id": 128001,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
11
  "initializer_range": 0.02,
@@ -20,15 +21,15 @@
20
  "rms_norm_eps": 1e-05,
21
  "rope_scaling": {
22
  "factor": 8.0,
23
- "low_freq_factor": 1.0,
24
  "high_freq_factor": 4.0,
 
25
  "original_max_position_embeddings": 8192,
26
  "rope_type": "llama3"
27
  },
28
  "rope_theta": 500000.0,
29
  "tie_word_embeddings": false,
30
  "torch_dtype": "bfloat16",
31
- "transformers_version": "4.43.0.dev0",
32
- "use_cache": true,
33
  "vocab_size": 128256
34
  }
 
1
  {
2
+ "_name_or_path": "teknium/Llama-3.1-AlternateTokenizer",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": 128009,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
 
21
  "rms_norm_eps": 1e-05,
22
  "rope_scaling": {
23
  "factor": 8.0,
 
24
  "high_freq_factor": 4.0,
25
+ "low_freq_factor": 1.0,
26
  "original_max_position_embeddings": 8192,
27
  "rope_type": "llama3"
28
  },
29
  "rope_theta": 500000.0,
30
  "tie_word_embeddings": false,
31
  "torch_dtype": "bfloat16",
32
+ "transformers_version": "4.44.0.dev0",
33
+ "use_cache": false,
34
  "vocab_size": 128256
35
  }
generation_config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
 
 
2
  "do_sample": true,
 
3
  "temperature": 0.6,
4
  "top_p": 0.9,
5
- "_from_model_config": true,
6
- "bos_token_id": 128000,
7
- "eos_token_id": 128001,
8
- "transformers_version": "4.43.0.dev0"
9
  }
 
1
  {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
  "do_sample": true,
5
+ "eos_token_id": 128009,
6
  "temperature": 0.6,
7
  "top_p": 0.9,
8
+ "transformers_version": "4.44.0.dev0"
 
 
 
9
  }
special_tokens_map.json CHANGED
@@ -1,4 +1,23 @@
1
  {
2
- "bos_token": "<|begin_of_text|>",
3
- "eos_token": "<|im_end|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|eot_id|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|end_of_text|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -313,20 +313,20 @@
313
  "special": false
314
  },
315
  "128039": {
316
- "content": "<|im_start|>",
317
  "lstrip": false,
318
  "normalized": false,
319
  "rstrip": false,
320
  "single_word": false,
321
- "special": false
322
  },
323
  "128040": {
324
- "content": "<|im_end|>",
325
  "lstrip": false,
326
  "normalized": false,
327
  "rstrip": false,
328
  "single_word": false,
329
- "special": true
330
  },
331
  "128041": {
332
  "content": "<|reserved_special_token_33|>",
@@ -2050,12 +2050,14 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
 
2053
  "clean_up_tokenization_spaces": true,
2054
- "eos_token": "<|end_of_text|>",
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
  "model_max_length": 131072,
 
2060
  "tokenizer_class": "PreTrainedTokenizerFast"
2061
  }
 
313
  "special": false
314
  },
315
  "128039": {
316
+ "content": "<|im_end|>",
317
  "lstrip": false,
318
  "normalized": false,
319
  "rstrip": false,
320
  "single_word": false,
321
+ "special": true
322
  },
323
  "128040": {
324
+ "content": "<|im_start|>",
325
  "lstrip": false,
326
  "normalized": false,
327
  "rstrip": false,
328
  "single_word": false,
329
+ "special": false
330
  },
331
  "128041": {
332
  "content": "<|reserved_special_token_33|>",
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|eot_id|>",
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
+ "pad_token": "<|end_of_text|>",
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }