martinkonopka
commited on
Commit
•
d9b0c33
1
Parent(s):
0632ee2
Add new SentenceTransformer model.
Browse files- .gitattributes +1 -0
- config.json +1 -1
- pytorch_model.bin +2 -2
- tokenizer_config.json +1 -1
.gitattributes
CHANGED
@@ -27,3 +27,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
28 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
29 |
.git/lfs/objects/7b/1c/7b1c3ca65da2275e74cc85ad22e76e66ecc6368914f0e9667029d1af967ad7ee filter=lfs diff=lfs merge=lfs -text
|
|
|
|
27 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
28 |
pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
29 |
.git/lfs/objects/7b/1c/7b1c3ca65da2275e74cc85ad22e76e66ecc6368914f0e9667029d1af967ad7ee filter=lfs diff=lfs merge=lfs -text
|
30 |
+
.git/lfs/objects/e1/f6/e1f6761555731e9e8e29ec3342163f2bbbbb3edd3395993ff3f57de27d0fa05b filter=lfs diff=lfs merge=lfs -text
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "models/sts/",
|
3 |
"architectures": [
|
4 |
"RobertaModel"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "models/sts/0_Transformer",
|
3 |
"architectures": [
|
4 |
"RobertaModel"
|
5 |
],
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c37550b2742afa71a1d36ec08b8c819f213562b045e25517b83e106f8f8c27f
|
3 |
+
size 498660807
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "errors": "replace", "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "special_tokens_map_file": null, "model_max_length": 512, "name_or_path": "models/sts/", "tokenizer_class": "RobertaTokenizer"}
|
|
|
1 |
+
{"unk_token": {"content": "<unk>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "errors": "replace", "sep_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "cls_token": {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "special_tokens_map_file": null, "model_max_length": 512, "name_or_path": "models/sts/0_Transformer", "tokenizer_class": "RobertaTokenizer"}
|