96abhishekarora commited on
Commit
3c4ad15
1 Parent(s): 3d73ad8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -11,12 +11,12 @@ tags:
11
  - news
12
  ---
13
 
14
- This model was contrastively trained for entity coreference on a dataset constructed by mentions of the same entity. The model requires text with entities detected via NER and focuses specifically on Person [PER] tags. The mention of an entity of interest is then decorated a set of special tokens with the syntax - ([M ] Entity [\M ] .
15
 
16
  We start with a base S-BERT MPNet bi-encoder [model](https://huggingface.co/sentence-transformers/all-mpnet-base-v2). This is contrastively trained on 179 million pairs taken from mentions of entities on Wikipedia, where positives are mentions of the same individual. Hard negatives are mined using individuals that appear on the same disambiguation pages.
17
  This essentially trains a semantic similarity model.
18
 
19
- At inference time, to collect mentions of the same entity, we reccomend a clustering contexts with deocrated entities in the embedding space - each cluster is then a set of mentions of the same entity, clusterd through the contextual information contained in the enclosing text.
20
 
21
 
22
  More information about its training (and use) can be found on the associated code [repo](https://github.com/dell-research-harvard/newswire/tree/main) and [paper](https://arxiv.org/pdf/2406.09490).
 
11
  - news
12
  ---
13
 
14
+ This model was contrastively trained for entity coreference on a dataset constructed by mentions of the same entity. The model requires text with entities detected via NER and focuses specifically on Person [PER] tags. The mention of an entity of interest is then decorated with a set of special tokens with the syntax - ([M ] Entity [\M ] .
15
 
16
  We start with a base S-BERT MPNet bi-encoder [model](https://huggingface.co/sentence-transformers/all-mpnet-base-v2). This is contrastively trained on 179 million pairs taken from mentions of entities on Wikipedia, where positives are mentions of the same individual. Hard negatives are mined using individuals that appear on the same disambiguation pages.
17
  This essentially trains a semantic similarity model.
18
 
19
+ At inference time, to collect mentions of the same entity, we recommend clustering contexts with decorated entities in the embedding space - each cluster is then a set of mentions of the same entity, clustered through the contextual information contained in the enclosing text.
20
 
21
 
22
  More information about its training (and use) can be found on the associated code [repo](https://github.com/dell-research-harvard/newswire/tree/main) and [paper](https://arxiv.org/pdf/2406.09490).