mmoreirast commited on
Commit
91cad6c
1 Parent(s): eb5e5e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: apache-2.0
3
  datasets:
4
- - mmoreirast/medicine-training-pt
5
  - mmoreirast/medicine-evaluation-pt
 
6
  language:
7
  - pt
8
  metrics:
@@ -31,7 +31,7 @@ You can check the codes used to fine-tune the model at the following [Google Col
31
  ## Fine-tuning details
32
  - **Base model:** [TeenyTinyLlama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m)
33
  - **Context length:** 2048 tokens
34
- - **Dataset for fine-tuning:** [medicine-training-pt](mmoreirast/medicine-training-pt)
35
  - **Dataset for evaluation:** [medicine-evaluation-pt](https://huggingface.co/datasets/mmoreirast/medicine-evaluation-pt)
36
  - **Language:** Portuguese
37
  - **GPU:** NVIDIA A100-SXM4-40GB
@@ -39,7 +39,7 @@ You can check the codes used to fine-tune the model at the following [Google Col
39
 
40
  ## Parameters
41
  - **Number of Epochs:** 4
42
- - **Batch size:** 8
43
  - **Optimizer:** torch.optim.AdamW (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
44
 
45
  ## Evaluations
@@ -61,7 +61,7 @@ Using the `pipeline`:
61
  ```python
62
  from transformers import pipeline
63
 
64
- generator = pipeline("text-generation", model="mmoreirast/Doctor-Llama-460m")
65
 
66
  completions = generator("Me fale sobre o sistema nervoso", num_return_sequences=2, max_new_tokens=100)
67
 
@@ -76,8 +76,8 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
76
  import torch
77
 
78
  # Load model and the tokenizer
79
- tokenizer = AutoTokenizer.from_pretrained("mmoreirast/Doctor-Llama-460m", revision='main')
80
- model = AutoModelForCausalLM.from_pretrained("mmoreirast/Doctor-Llama-460m", revision='main')
81
 
82
  # Pass the model to your device
83
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
 
1
  ---
2
  license: apache-2.0
3
  datasets:
 
4
  - mmoreirast/medicine-evaluation-pt
5
+ - mmoreirast/aira-med-training-pt
6
  language:
7
  - pt
8
  metrics:
 
31
  ## Fine-tuning details
32
  - **Base model:** [TeenyTinyLlama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m)
33
  - **Context length:** 2048 tokens
34
+ - **Dataset for fine-tuning:** [aira-med-training-pt](https://huggingface.co/datasets/mmoreirast/aira-med-training-pt)
35
  - **Dataset for evaluation:** [medicine-evaluation-pt](https://huggingface.co/datasets/mmoreirast/medicine-evaluation-pt)
36
  - **Language:** Portuguese
37
  - **GPU:** NVIDIA A100-SXM4-40GB
 
39
 
40
  ## Parameters
41
  - **Number of Epochs:** 4
42
+ - **Batch size:** 3
43
  - **Optimizer:** torch.optim.AdamW (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
44
 
45
  ## Evaluations
 
61
  ```python
62
  from transformers import pipeline
63
 
64
+ generator = pipeline("text-generation", model="mmoreirast/Doctor-Llama-Chat")
65
 
66
  completions = generator("Me fale sobre o sistema nervoso", num_return_sequences=2, max_new_tokens=100)
67
 
 
76
  import torch
77
 
78
  # Load model and the tokenizer
79
+ tokenizer = AutoTokenizer.from_pretrained("mmoreirast/Doctor-Llama-Chat", revision='main')
80
+ model = AutoModelForCausalLM.from_pretrained("mmoreirast/Doctor-Llama-Chat", revision='main')
81
 
82
  # Pass the model to your device
83
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")