sooolee commited on
Commit
546e33f
·
1 Parent(s): a195413

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -27,7 +27,11 @@ inference: false
27
 
28
  # bart-large-cnn-finetuned-samsum-lora
29
 
30
- This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset. The base model [bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) is a fine-tuned verstion of BART model on the [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail) dataset.
 
 
 
 
31
 
32
  ## Model description
33
 
@@ -62,7 +66,7 @@ Input texts longer than 4000 words can be converted into a list of sub-texts wit
62
  Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
63
  Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
64
 
65
- Note 'max_new_tokens=60' is used to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
66
 
67
  ```python
68
  import torch
 
27
 
28
  # bart-large-cnn-finetuned-samsum-lora
29
 
30
+ This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset.
31
+
32
+ The base model [bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) is a fine-tuned verstion of BART model on the [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail) dataset.
33
+
34
+ Check out [sooolee/flan-t5-base-cnn-samsum-lora](https://huggingface.co/sooolee/flan-t5-base-cnn-samsum-lora) the model fine-tuned for the same purpose.
35
 
36
  ## Model description
37
 
 
66
  Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
67
  Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
68
 
69
+ Note 'max_new_tokens=60' is used in the example below to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
70
 
71
  ```python
72
  import torch