mmoreirast
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
datasets:
|
4 |
-
- mmoreirast/medicine-training-pt
|
5 |
- mmoreirast/medicine-evaluation-pt
|
|
|
6 |
language:
|
7 |
- pt
|
8 |
metrics:
|
@@ -31,7 +31,7 @@ You can check the codes used to fine-tune the model at the following [Google Col
|
|
31 |
## Fine-tuning details
|
32 |
- **Base model:** [TeenyTinyLlama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m)
|
33 |
- **Context length:** 2048 tokens
|
34 |
-
- **Dataset for fine-tuning:** [
|
35 |
- **Dataset for evaluation:** [medicine-evaluation-pt](https://huggingface.co/datasets/mmoreirast/medicine-evaluation-pt)
|
36 |
- **Language:** Portuguese
|
37 |
- **GPU:** NVIDIA A100-SXM4-40GB
|
@@ -39,7 +39,7 @@ You can check the codes used to fine-tune the model at the following [Google Col
|
|
39 |
|
40 |
## Parameters
|
41 |
- **Number of Epochs:** 4
|
42 |
-
- **Batch size:**
|
43 |
- **Optimizer:** torch.optim.AdamW (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
|
44 |
|
45 |
## Evaluations
|
@@ -61,7 +61,7 @@ Using the `pipeline`:
|
|
61 |
```python
|
62 |
from transformers import pipeline
|
63 |
|
64 |
-
generator = pipeline("text-generation", model="mmoreirast/Doctor-Llama-
|
65 |
|
66 |
completions = generator("Me fale sobre o sistema nervoso", num_return_sequences=2, max_new_tokens=100)
|
67 |
|
@@ -76,8 +76,8 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
76 |
import torch
|
77 |
|
78 |
# Load model and the tokenizer
|
79 |
-
tokenizer = AutoTokenizer.from_pretrained("mmoreirast/Doctor-Llama-
|
80 |
-
model = AutoModelForCausalLM.from_pretrained("mmoreirast/Doctor-Llama-
|
81 |
|
82 |
# Pass the model to your device
|
83 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
datasets:
|
|
|
4 |
- mmoreirast/medicine-evaluation-pt
|
5 |
+
- mmoreirast/aira-med-training-pt
|
6 |
language:
|
7 |
- pt
|
8 |
metrics:
|
|
|
31 |
## Fine-tuning details
|
32 |
- **Base model:** [TeenyTinyLlama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m)
|
33 |
- **Context length:** 2048 tokens
|
34 |
+
- **Dataset for fine-tuning:** [aira-med-training-pt](https://huggingface.co/datasets/mmoreirast/aira-med-training-pt)
|
35 |
- **Dataset for evaluation:** [medicine-evaluation-pt](https://huggingface.co/datasets/mmoreirast/medicine-evaluation-pt)
|
36 |
- **Language:** Portuguese
|
37 |
- **GPU:** NVIDIA A100-SXM4-40GB
|
|
|
39 |
|
40 |
## Parameters
|
41 |
- **Number of Epochs:** 4
|
42 |
+
- **Batch size:** 3
|
43 |
- **Optimizer:** torch.optim.AdamW (warmup_steps = 1e3, learning_rate = 1e-5, epsilon = 1e-8)
|
44 |
|
45 |
## Evaluations
|
|
|
61 |
```python
|
62 |
from transformers import pipeline
|
63 |
|
64 |
+
generator = pipeline("text-generation", model="mmoreirast/Doctor-Llama-Chat")
|
65 |
|
66 |
completions = generator("Me fale sobre o sistema nervoso", num_return_sequences=2, max_new_tokens=100)
|
67 |
|
|
|
76 |
import torch
|
77 |
|
78 |
# Load model and the tokenizer
|
79 |
+
tokenizer = AutoTokenizer.from_pretrained("mmoreirast/Doctor-Llama-Chat", revision='main')
|
80 |
+
model = AutoModelForCausalLM.from_pretrained("mmoreirast/Doctor-Llama-Chat", revision='main')
|
81 |
|
82 |
# Pass the model to your device
|
83 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|