Philip May
commited on
Commit
•
31c912f
1
Parent(s):
2fb51bf
Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ metrics:
|
|
19 |
---
|
20 |
|
21 |
# Cross English & German RoBERTa for Sentence Embeddings
|
22 |
-
This model is intended to [compute sentence (text) embeddings](https://www.sbert.net/
|
23 |
|
24 |
The speciality of this model is that it also works cross-lingually. Regardless of the language, the sentences are translated into very similar vectors according to their semantics. This means that you can, for example, enter a search in German and find results according to the semantics in German and also in English. Using a xlm model and _multilingual finetuning with language-crossing_ we reach performance that even exceeds the best current dedicated English large model (see Evaluation section below).
|
25 |
|
|
|
19 |
---
|
20 |
|
21 |
# Cross English & German RoBERTa for Sentence Embeddings
|
22 |
+
This model is intended to [compute sentence (text) embeddings](https://www.sbert.net/examples/applications/computing-embeddings/README.html) for English and German text. These embeddings can then be compared with [cosine-similarity](https://en.wikipedia.org/wiki/Cosine_similarity) to find sentences with a similar semantic meaning. For example this can be useful for [semantic textual similarity](https://www.sbert.net/docs/usage/semantic_textual_similarity.html), [semantic search](https://www.sbert.net/docs/usage/semantic_search.html), or [paraphrase mining](https://www.sbert.net/docs/usage/paraphrase_mining.html). To do this you have to use the [Sentence Transformers Python framework](https://github.com/UKPLab/sentence-transformers).
|
23 |
|
24 |
The speciality of this model is that it also works cross-lingually. Regardless of the language, the sentences are translated into very similar vectors according to their semantics. This means that you can, for example, enter a search in German and find results according to the semantics in German and also in English. Using a xlm model and _multilingual finetuning with language-crossing_ we reach performance that even exceeds the best current dedicated English large model (see Evaluation section below).
|
25 |
|