Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,36 @@
|
|
1 |
---
|
2 |
license: llama3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: llama3
|
3 |
+
datasets:
|
4 |
+
- Henrychur/MMedC
|
5 |
+
- Henrychur/MedS-Ins
|
6 |
+
language:
|
7 |
+
- en
|
8 |
+
base_model: Henrychur/MMedS-Llama-3-8B
|
9 |
+
tags:
|
10 |
+
- medical
|
11 |
---
|
12 |
+
|
13 |
+
|
14 |
+
# MMedS-Llama3
|
15 |
+
[💻Github Repo](https://github.com/MAGIC-AI4Med/MedS-Ins) [🖨️arXiv Paper](https://arxiv.org/abs/2408.12547)
|
16 |
+
|
17 |
+
The official codes for "Towards Evaluating and Building Versatile Large Language Models for Medicine"
|
18 |
+
|
19 |
+
|
20 |
+
## Introduction
|
21 |
+
This repository hosts MMedS-Llama-3-8B. Its foundation model, [MMed-Llama-3-8B](https://huggingface.co/Henrychur/MMed-Llama-3-8B),
|
22 |
+
is a multilingual medical language model which has undergone additional continuous pretraining on MMedC. Furthermore, the model has
|
23 |
+
been fine-tuned under supervision using MedS-Ins, a comprehensive dataset designed specifically for supervised fine-tuning (SFT),
|
24 |
+
featuring 13.5 million samples across 122 tasks. For more details, please refer to our paper.
|
25 |
+
|
26 |
+
|
27 |
+
## Usage
|
28 |
+
The model can be loaded as follows:
|
29 |
+
```py
|
30 |
+
import torch
|
31 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
32 |
+
tokenizer = AutoTokenizer.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns")
|
33 |
+
model = AutoModelForCausalLM.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns", torch_dtype=torch.float16)
|
34 |
+
```
|
35 |
+
|
36 |
+
- Inference format is the same as Llama 3, you can check the inference code [here](https://github.com/MAGIC-AI4Med/MedS-Ins/blob/main/Inference/model.py).
|