Aansh123 commited on
Commit
c2a0d9f
1 Parent(s): ee8fb69

Training in progress, step 500

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b905fc5abbdfd3e1fc9fef338fc1ac02827f54b7968332dd15eaad67485b89af
3
  size 433327921
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abe30e1954c4c01953f21a9be62da7eeacfc2a08d57ff622e6de0783dc5df2ec
3
  size 433327921
runs/Aug16_12-32-15_8661ae855cce/events.out.tfevents.1692199272.8661ae855cce.418.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a6613c36694db07d19fa240e578072b3a5da5dd6877b8d8b02846166974d0c
3
+ size 4382
runs/Aug16_12-32-15_8661ae855cce/events.out.tfevents.1692199308.8661ae855cce.418.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fec69bb9752304691c5e2375e886308c31541f7217b7c1655d6cdc27be1506db
3
+ size 4862
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "BertTokenizer",
12
+ "unk_token": "[UNK]"
13
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff