Update README.md
Browse files
README.md
CHANGED
@@ -27,7 +27,11 @@ inference: false
|
|
27 |
|
28 |
# bart-large-cnn-finetuned-samsum-lora
|
29 |
|
30 |
-
This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset.
|
|
|
|
|
|
|
|
|
31 |
|
32 |
## Model description
|
33 |
|
@@ -62,7 +66,7 @@ Input texts longer than 4000 words can be converted into a list of sub-texts wit
|
|
62 |
Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
|
63 |
Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
|
64 |
|
65 |
-
Note 'max_new_tokens=60' is used to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
|
66 |
|
67 |
```python
|
68 |
import torch
|
|
|
27 |
|
28 |
# bart-large-cnn-finetuned-samsum-lora
|
29 |
|
30 |
+
This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset.
|
31 |
+
|
32 |
+
The base model [bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) is a fine-tuned verstion of BART model on the [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail) dataset.
|
33 |
+
|
34 |
+
Check out [sooolee/flan-t5-base-cnn-samsum-lora](https://huggingface.co/sooolee/flan-t5-base-cnn-samsum-lora) the model fine-tuned for the same purpose.
|
35 |
|
36 |
## Model description
|
37 |
|
|
|
66 |
Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
|
67 |
Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
|
68 |
|
69 |
+
Note 'max_new_tokens=60' is used in the example below to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
|
70 |
|
71 |
```python
|
72 |
import torch
|