File size: 1,100 Bytes
9d84f0b
 
 
 
 
0c2eff5
9d84f0b
0c2eff5
 
 
 
 
 
 
 
 
 
 
517a101
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
---
language: 
  - pt
---

This model was distilled from [BERTimbau](https://huggingface.co/neuralmind/bert-base-portuguese-cased)

## Usage

```python
from transformers import AutoTokenizer  # Or BertTokenizer
from transformers import AutoModelForPreTraining  # Or BertForPreTraining for loading pretraining heads
from transformers import AutoModel  # or BertModel, for BERT without pretraining heads
model = AutoModelForPreTraining.from_pretrained('adalbertojunior/distilbert-portuguese-cased')
tokenizer = AutoTokenizer.from_pretrained('adalbertojunior/distilbert-portuguese-cased', do_lower_case=False)
```
You should fine tune it on your own data.

It can achieve accuracy up to 99% relative to the original BERTimbau in some tasks.
```
@misc {adalberto_ferreira_barbosa_junior_2024,
	author       = { {Adalberto Ferreira Barbosa Junior} },
	title        = { distilbert-portuguese-cased (Revision df1fa7a) },
	year         = 2024,
	url          = { https://huggingface.co/adalbertojunior/distilbert-portuguese-cased },
	doi          = { 10.57967/hf/3041 },
	publisher    = { Hugging Face }
}
```