Fix paper link
Browse files
README.md
CHANGED
@@ -14,7 +14,7 @@ ByT5 was only pre-trained on [mC4](https://www.tensorflow.org/datasets/catalog/c
|
|
14 |
|
15 |
ByT5 works especially well on noisy text data,*e.g.*, `google/byt5-large` significantly outperforms [mt5-large](https://huggingface.co/google/mt5-large) on [TweetQA](https://arxiv.org/abs/1907.06292).
|
16 |
|
17 |
-
Paper: [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/
|
18 |
|
19 |
Authors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel*
|
20 |
|
|
|
14 |
|
15 |
ByT5 works especially well on noisy text data,*e.g.*, `google/byt5-large` significantly outperforms [mt5-large](https://huggingface.co/google/mt5-large) on [TweetQA](https://arxiv.org/abs/1907.06292).
|
16 |
|
17 |
+
Paper: [ByT5: Towards a token-free future with pre-trained byte-to-byte models](https://arxiv.org/abs/2105.13626)
|
18 |
|
19 |
Authors: *Linting Xue, Aditya Barua, Noah Constant, Rami Al-Rfou, Sharan Narang, Mihir Kale, Adam Roberts, Colin Raffel*
|
20 |
|