|
--- |
|
language: |
|
- ko |
|
license: cc-by-nc-2.0 |
|
size_categories: |
|
- 10K<n<100K |
|
task_categories: |
|
- question-answering |
|
configs: |
|
- config_name: dentist |
|
data_files: |
|
- split: train |
|
path: dentist/train-* |
|
- split: dev |
|
path: dentist/dev-* |
|
- split: test |
|
path: dentist/test-* |
|
- split: fewshot |
|
path: dentist/fewshot-* |
|
- config_name: doctor |
|
data_files: |
|
- split: train |
|
path: doctor/train-* |
|
- split: dev |
|
path: doctor/dev-* |
|
- split: test |
|
path: doctor/test-* |
|
- split: fewshot |
|
path: doctor/fewshot-* |
|
- config_name: nurse |
|
data_files: |
|
- split: train |
|
path: nurse/train-* |
|
- split: dev |
|
path: nurse/dev-* |
|
- split: test |
|
path: nurse/test-* |
|
- split: fewshot |
|
path: nurse/fewshot-* |
|
- config_name: pharm |
|
data_files: |
|
- split: train |
|
path: pharm/train-* |
|
- split: dev |
|
path: pharm/dev-* |
|
- split: test |
|
path: pharm/test-* |
|
- split: fewshot |
|
path: pharm/fewshot-* |
|
tags: |
|
- medical |
|
dataset_info: |
|
- config_name: dentist |
|
features: |
|
- name: subject |
|
dtype: string |
|
- name: year |
|
dtype: int64 |
|
- name: period |
|
dtype: int64 |
|
- name: q_number |
|
dtype: int64 |
|
- name: question |
|
dtype: string |
|
- name: A |
|
dtype: string |
|
- name: B |
|
dtype: string |
|
- name: C |
|
dtype: string |
|
- name: D |
|
dtype: string |
|
- name: E |
|
dtype: string |
|
- name: answer |
|
dtype: int64 |
|
- name: cot |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 116376 |
|
num_examples: 297 |
|
- name: dev |
|
num_bytes: 119727 |
|
num_examples: 304 |
|
- name: test |
|
num_bytes: 330325 |
|
num_examples: 811 |
|
- name: fewshot |
|
num_bytes: 4810 |
|
num_examples: 5 |
|
download_size: 374097 |
|
dataset_size: 571238 |
|
- config_name: doctor |
|
features: |
|
- name: subject |
|
dtype: string |
|
- name: year |
|
dtype: int64 |
|
- name: period |
|
dtype: int64 |
|
- name: q_number |
|
dtype: int64 |
|
- name: question |
|
dtype: string |
|
- name: A |
|
dtype: string |
|
- name: B |
|
dtype: string |
|
- name: C |
|
dtype: string |
|
- name: D |
|
dtype: string |
|
- name: E |
|
dtype: string |
|
- name: answer |
|
dtype: int64 |
|
- name: cot |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 1137189 |
|
num_examples: 1890 |
|
- name: dev |
|
num_bytes: 111294 |
|
num_examples: 164 |
|
- name: test |
|
num_bytes: 315104 |
|
num_examples: 435 |
|
- name: fewshot |
|
num_bytes: 8566 |
|
num_examples: 5 |
|
download_size: 871530 |
|
dataset_size: 1572153 |
|
- config_name: nurse |
|
features: |
|
- name: subject |
|
dtype: string |
|
- name: year |
|
dtype: int64 |
|
- name: period |
|
dtype: int64 |
|
- name: q_number |
|
dtype: int64 |
|
- name: question |
|
dtype: string |
|
- name: A |
|
dtype: string |
|
- name: B |
|
dtype: string |
|
- name: C |
|
dtype: string |
|
- name: D |
|
dtype: string |
|
- name: E |
|
dtype: string |
|
- name: answer |
|
dtype: int64 |
|
- name: cot |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 219983 |
|
num_examples: 582 |
|
- name: dev |
|
num_bytes: 110210 |
|
num_examples: 291 |
|
- name: test |
|
num_bytes: 327186 |
|
num_examples: 878 |
|
- name: fewshot |
|
num_bytes: 6324 |
|
num_examples: 5 |
|
download_size: 419872 |
|
dataset_size: 663703 |
|
- config_name: pharm |
|
features: |
|
- name: subject |
|
dtype: string |
|
- name: year |
|
dtype: int64 |
|
- name: period |
|
dtype: int64 |
|
- name: q_number |
|
dtype: int64 |
|
- name: question |
|
dtype: string |
|
- name: A |
|
dtype: string |
|
- name: B |
|
dtype: string |
|
- name: C |
|
dtype: string |
|
- name: D |
|
dtype: string |
|
- name: E |
|
dtype: string |
|
- name: answer |
|
dtype: int64 |
|
- name: cot |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 272256 |
|
num_examples: 632 |
|
- name: dev |
|
num_bytes: 139900 |
|
num_examples: 300 |
|
- name: test |
|
num_bytes: 412847 |
|
num_examples: 885 |
|
- name: fewshot |
|
num_bytes: 6324 |
|
num_examples: 5 |
|
download_size: 504010 |
|
dataset_size: 831327 |
|
--- |
|
|
|
# KorMedMCQA : Multi-Choice Question Answering Benchmark for Korean Healthcare Professional Licensing Examinations |
|
|
|
We present KorMedMCQA, the first Korean Medical Multiple-Choice Question |
|
Answering benchmark, derived from professional healthcare licensing |
|
examinations conducted in Korea between 2012 and 2024. The dataset contains |
|
7,469 questions from examinations for doctor, nurse, pharmacist, and dentist, |
|
covering a wide range of medical disciplines. We evaluate the performance of 59 |
|
large language models, spanning proprietary and open-source models, |
|
multilingual and Korean-specialized models, and those fine-tuned for clinical |
|
applications. Our results show that applying Chain of Thought (CoT) reasoning |
|
can enhance the model performance by up to 4.5% compared to direct answering |
|
approaches. We also investigate whether MedQA, one of the most widely used |
|
medical benchmarks derived from the U.S. Medical Licensing Examination, can |
|
serve as a reliable proxy for evaluating model performance in other regions-in |
|
this case, Korea. Our correlation analysis between model scores on KorMedMCQA |
|
and MedQA reveals that these two benchmarks align no better than benchmarks |
|
from entirely different domains (e.g., MedQA and MMLU-Pro). This finding |
|
underscores the substantial linguistic and clinical differences between Korean |
|
and U.S. medical contexts, reinforcing the need for region-specific medical QA |
|
benchmarks. |
|
|
|
Paper : https://arxiv.org/abs/2403.01469 |
|
|
|
## Notice |
|
|
|
We have made the following updates to the KorMedMCQA dataset: |
|
|
|
1. **Dentist Exam**: Incorporated exam questions from 2021 to 2024. |
|
2. **Updated Test Sets**: Added the 2024 exam questions for the doctor, nurse, and pharmacist test sets. |
|
3. **Few-Shot Split**: Introduced a `fewshot` split, containing 5 shots from each validation set. |
|
4. **Chain-of-Thought(CoT)**: In each exam's few-shot split (`cot` column), there is an answer with reasoning annotated by professionals |
|
|
|
|
|
## Dataset Details |
|
|
|
### Languages |
|
|
|
Korean |
|
|
|
### Subtask |
|
|
|
``` |
|
from datasets import load_dataset |
|
doctor = load_dataset(path = "sean0042/KorMedMCQA",name = "doctor") |
|
nurse = load_dataset(path = "sean0042/KorMedMCQA",name = "nurse") |
|
pharmacist = load_dataset(path = "sean0042/KorMedMCQA",name = "pharm") |
|
dentist = load_dataset(path = "sean0042/KorMedMCQA",name = "dentist") |
|
``` |
|
|
|
### Statistics |
|
|
|
| Category | # Questions (Train/Dev/Test) | |
|
|------------------------------|------------------------------| |
|
| Doctor | 2,489 (1,890/164/435) | |
|
| Nurse | 1,751 (582/291/878) | |
|
| Pharmacist | 1,817 (632/300/885) | |
|
| Dentist | 1,412 (297/304/811) | |
|
|
|
### Data Fields |
|
|
|
|
|
- `subject`: doctor, nurse, or pharm |
|
- `year`: year of the examination |
|
- `period`: period of the examination |
|
- `q_number`: question number of the examination |
|
- `question`: question |
|
- `A`: First answer choice |
|
- `B`: Second answer choice |
|
- `C`: Third answer choice |
|
- `D`: Fourth answer choice |
|
- `E`: Fifth answer choice |
|
- `cot` : Answer with reasoning annotated by professionals (only available in fewshot split) |
|
- `answer` : Answer (1 to 5). 1 denotes answer A, and 5 denotes answer E |
|
|
|
|
|
## Contact |
|
|
|
``` |
|
sean0042@kaist.ac.kr |
|
``` |