File size: 1,100 Bytes
9d84f0b 0c2eff5 9d84f0b 0c2eff5 517a101 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 |
---
language:
- pt
---
This model was distilled from [BERTimbau](https://huggingface.co/neuralmind/bert-base-portuguese-cased)
## Usage
```python
from transformers import AutoTokenizer # Or BertTokenizer
from transformers import AutoModelForPreTraining # Or BertForPreTraining for loading pretraining heads
from transformers import AutoModel # or BertModel, for BERT without pretraining heads
model = AutoModelForPreTraining.from_pretrained('adalbertojunior/distilbert-portuguese-cased')
tokenizer = AutoTokenizer.from_pretrained('adalbertojunior/distilbert-portuguese-cased', do_lower_case=False)
```
You should fine tune it on your own data.
It can achieve accuracy up to 99% relative to the original BERTimbau in some tasks.
```
@misc {adalberto_ferreira_barbosa_junior_2024,
author = { {Adalberto Ferreira Barbosa Junior} },
title = { distilbert-portuguese-cased (Revision df1fa7a) },
year = 2024,
url = { https://huggingface.co/adalbertojunior/distilbert-portuguese-cased },
doi = { 10.57967/hf/3041 },
publisher = { Hugging Face }
}
``` |