ww
commited on
Commit
路
a52378e
1
Parent(s):
ab4ff29
README.md
CHANGED
@@ -31,17 +31,16 @@ model-index:
|
|
31 |
verified: true
|
32 |
verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFkMmI2ODM0NmY5NGNkNmUxYWViOWYxZDNkY2EzYWFmOWI4N2VhYzY5MGEzMTVhOTU4Zjc4YWViOGNjOWJjMCIsInZlcnNpb24iOjF9.fexrU1icJK5_MiifBtZWkeUvpmFISqBLDXSQJ8E6UnrRof-7cU0s4tX_dIsauHWtUpIHMPZCf5dlMWQKXZuAAA
|
33 |
---
|
|
|
34 |
|
35 |
-
|
36 |
-
|
37 |
-
This is the best model for QA [roberta-base](https://huggingface.co/roberta-base) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
|
38 |
-
|
39 |
|
40 |
## Overview
|
41 |
-
|
42 |
-
**Language:**
|
43 |
-
**
|
44 |
-
**
|
|
|
45 |
**Eval data:** SQuAD 2.0
|
46 |
**Infrastructure**: 4x Tesla v100
|
47 |
|
@@ -57,20 +56,24 @@ lr_schedule = LinearWarmup
|
|
57 |
warmup_proportion = 0.2
|
58 |
doc_stride=128
|
59 |
max_query_length=64
|
60 |
-
```
|
61 |
|
62 |
## Usage
|
63 |
|
64 |
### In Haystack
|
|
|
65 |
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
|
|
|
66 |
```python
|
67 |
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
|
68 |
# or
|
69 |
reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
|
70 |
```
|
|
|
71 |
For a complete example of ``roberta-base-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
|
72 |
|
73 |
### In Transformers
|
|
|
74 |
```python
|
75 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
|
76 |
|
@@ -90,6 +93,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
90 |
```
|
91 |
|
92 |
## Performance
|
|
|
93 |
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
|
94 |
|
95 |
```
|
|
|
31 |
verified: true
|
32 |
verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFkMmI2ODM0NmY5NGNkNmUxYWViOWYxZDNkY2EzYWFmOWI4N2VhYzY5MGEzMTVhOTU4Zjc4YWViOGNjOWJjMCIsInZlcnNpb24iOjF9.fexrU1icJK5_MiifBtZWkeUvpmFISqBLDXSQJ8E6UnrRof-7cU0s4tX_dIsauHWtUpIHMPZCf5dlMWQKXZuAAA
|
33 |
---
|
34 |
+
# algmon-base for QA
|
35 |
|
36 |
+
This is the base model for QA [roberta-base](https://huggingface.co/roberta-base) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
|
|
|
|
|
|
|
37 |
|
38 |
## Overview
|
39 |
+
|
40 |
+
**Language model:** roberta-base
|
41 |
+
**Language:** English
|
42 |
+
**Downstream-task:** Extractive QA
|
43 |
+
**Training data:** SQuAD 2.0
|
44 |
**Eval data:** SQuAD 2.0
|
45 |
**Infrastructure**: 4x Tesla v100
|
46 |
|
|
|
56 |
warmup_proportion = 0.2
|
57 |
doc_stride=128
|
58 |
max_query_length=64
|
59 |
+
```
|
60 |
|
61 |
## Usage
|
62 |
|
63 |
### In Haystack
|
64 |
+
|
65 |
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
|
66 |
+
|
67 |
```python
|
68 |
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
|
69 |
# or
|
70 |
reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
|
71 |
```
|
72 |
+
|
73 |
For a complete example of ``roberta-base-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
|
74 |
|
75 |
### In Transformers
|
76 |
+
|
77 |
```python
|
78 |
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
|
79 |
|
|
|
93 |
```
|
94 |
|
95 |
## Performance
|
96 |
+
|
97 |
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
|
98 |
|
99 |
```
|