Matthew Kenney commited on
Commit
4ef5385
1 Parent(s): 372598e

commit files to HF hub

Browse files
Files changed (4) hide show
  1. config.json +2 -2
  2. model.safetensors +2 -2
  3. tokenizer.json +2 -2
  4. tokenizer_config.json +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/content/results_16_epoch_1e_3/checkpoint-4005",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -16,7 +16,7 @@
16
  "n_head": 6,
17
  "n_inner": null,
18
  "n_layer": 6,
19
- "n_positions": 1024,
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
 
1
  {
2
+ "_name_or_path": "/content/vsmall_256_epoch_1e_3/checkpoint-420",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
16
  "n_head": 6,
17
  "n_inner": null,
18
  "n_layer": 6,
19
+ "n_positions": 256,
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec6fb7e21d8dcd22632aeb8e6f448981955291c9c47b0704e023a144fa11987d
3
- size 121365368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6f4326b673ed0ad7d96c5c40879e2af28d8a514650c14bf8d35280c46aa029
3
+ size 120185720
tokenizer.json CHANGED
@@ -2,13 +2,13 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
- "Fixed": 512
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
  "padding": {
10
  "strategy": {
11
+ "Fixed": 256
12
  },
13
  "direction": "Right",
14
  "pad_to_multiple_of": null,
tokenizer_config.json CHANGED
@@ -13,7 +13,7 @@
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
- "max_length": 512,
17
  "model_max_length": 1024,
18
  "pad_to_multiple_of": null,
19
  "pad_token": "<|endoftext|>",
 
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
+ "max_length": 256,
17
  "model_max_length": 1024,
18
  "pad_to_multiple_of": null,
19
  "pad_token": "<|endoftext|>",