Update README.md
Browse files
README.md
CHANGED
@@ -78,7 +78,7 @@ distilled_student_sentiment_classifier("私はこの映画が大好きで、何
|
|
78 |
|
79 |
## Training procedure
|
80 |
|
81 |
-
Notebook link:
|
82 |
|
83 |
### Training hyperparameters
|
84 |
|
@@ -99,19 +99,19 @@ python transformers/examples/research_projects/zero-shot-distillation/distill_cl
|
|
99 |
|
100 |
If you are training this model on Colab, make the following code changes to avoid Out-of-memory error message:
|
101 |
```bash
|
102 |
-
|
103 |
-
|
104 |
|
105 |
-
|
106 |
-
|
107 |
|
108 |
-
|
109 |
-
|
110 |
-
|
111 |
|
112 |
-
|
113 |
-
|
114 |
-
|
115 |
|
116 |
```
|
117 |
|
|
|
78 |
|
79 |
## Training procedure
|
80 |
|
81 |
+
Notebook link: [here](https://github.com/LxYuan0420/nlp/blob/main/notebooks/Distilling_Zero_Shot_multilingual_distilbert_sentiments_student.ipynb)
|
82 |
|
83 |
### Training hyperparameters
|
84 |
|
|
|
99 |
|
100 |
If you are training this model on Colab, make the following code changes to avoid Out-of-memory error message:
|
101 |
```bash
|
102 |
+
###### modify L78 to disable fast tokenizer
|
103 |
+
default=False,
|
104 |
|
105 |
+
###### update dataset map part at L313
|
106 |
+
dataset = dataset.map(tokenizer, input_columns="text", fn_kwargs={"padding": "max_length", "truncation": True, "max_length": 512})
|
107 |
|
108 |
+
###### add following lines to L213
|
109 |
+
del model
|
110 |
+
print(f"Manually deleted Teacher model, free some memory for student model.")
|
111 |
|
112 |
+
###### add following lines to L337
|
113 |
+
trainer.push_to_hub()
|
114 |
+
tokenizer.push_to_hub("distilbert-base-multilingual-cased-sentiments-student")
|
115 |
|
116 |
```
|
117 |
|