macadeliccc commited on
Commit
f00b978
1 Parent(s): 4dc531a

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "unsloth/llama-3-8b",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -21,9 +21,8 @@
21
  "rope_scaling": null,
22
  "rope_theta": 500000.0,
23
  "tie_word_embeddings": false,
24
- "torch_dtype": "bfloat16",
25
- "transformers_version": "4.40.0",
26
- "unsloth_version": "2024.4",
27
  "use_cache": true,
28
  "vocab_size": 128256
29
  }
 
1
  {
2
+ "_name_or_path": "meta-llama/Meta-Llama-3-8B",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
21
  "rope_scaling": null,
22
  "rope_theta": 500000.0,
23
  "tie_word_embeddings": false,
24
+ "torch_dtype": "float16",
25
+ "transformers_version": "4.40.0.dev0",
 
26
  "use_cache": true,
27
  "vocab_size": 128256
28
  }
generation_config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "_from_model_config": true,
3
  "bos_token_id": 128000,
4
- "eos_token_id": [
5
- 128001,
6
- 128009
7
- ],
8
- "transformers_version": "4.40.0"
 
9
  }
 
1
  {
 
2
  "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": 128001,
5
+ "max_length": 4096,
6
+ "temperature": 0.6,
7
+ "top_p": 0.9,
8
+ "transformers_version": "4.40.0.dev0"
9
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d512133eaea665a7499853f93ba2cd4ac31193235228edeb10c043b22ec4b4aa
3
- size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d9f4f21dd3cb8df5b8b82b3e1b46d004945fa70f7259a5a77fae6ec67a44478
3
+ size 4976698592
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e5c775f878b91c2d25b96ec752ad755c39994fff4d42be921224f2947611db4
3
- size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e01cace598750f03b75ba43fa657e2be4979ee5439c5beea3c57528e2785a1ad
3
+ size 4999802616
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d1c25295d29002ed752b3674ed280e3a4fb215bdb941ba79eadb875bfd7c21c
3
- size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96fe704a565e639dade8c90f7cc278746867890127542feb69d6d95ac6da167b
3
+ size 4915916080
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:963ce9b9174a206cf3f596d3e3ad0adac69f432d844a0419f2cf4a18925a8480
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03cba5c2d294dc2f012619cf3fc64a93dddc0e9cdb8e8bd45204ad3b3f287940
3
  size 1168138808
special_tokens_map.json CHANGED
@@ -12,12 +12,5 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|end_of_text|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
  }
23
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
 
 
 
 
 
 
 
15
  }
16
  }
tokenizer.json CHANGED
@@ -2407,7 +2407,6 @@
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
2410
- "ignore_merges": true,
2411
  "vocab": {
2412
  "!": 0,
2413
  "\"": 1,
 
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
 
2410
  "vocab": {
2411
  "!": 0,
2412
  "\"": 1,
tokenizer_config.json CHANGED
@@ -2056,8 +2056,6 @@
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
- "model_max_length": 8192,
2060
- "pad_token": "<|end_of_text|>",
2061
- "padding_side": "left",
2062
  "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }
 
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
+ "model_max_length": 1000000000000000019884624838656,
 
 
2060
  "tokenizer_class": "PreTrainedTokenizerFast"
2061
  }