Update README.md
Browse filesUpdate link to distillation process
README.md
CHANGED
@@ -11,7 +11,7 @@ datasets:
|
|
11 |
This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-cased).
|
12 |
It was introduced in [this paper](https://arxiv.org/abs/1910.01108).
|
13 |
The code for the distillation process can be found
|
14 |
-
[here](https://github.com/huggingface/transformers/tree/
|
15 |
This model is cased: it does make a difference between english and English.
|
16 |
|
17 |
All the training details on the pre-training, the uses, limitations and potential biases (included below) are the same as for [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased).
|
|
|
11 |
This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-cased).
|
12 |
It was introduced in [this paper](https://arxiv.org/abs/1910.01108).
|
13 |
The code for the distillation process can be found
|
14 |
+
[here](https://github.com/huggingface/transformers/tree/main/examples/research_projects/distillation).
|
15 |
This model is cased: it does make a difference between english and English.
|
16 |
|
17 |
All the training details on the pre-training, the uses, limitations and potential biases (included below) are the same as for [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased).
|