egrace479 commited on
Commit
7509edd
·
1 Parent(s): 278d188

Update README.md

Browse files

fixed and added links

Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -18,9 +18,9 @@ This model takes in an image of a fish and segments out traits, as described [be
18
  `se_resnext50_32x4d-a260b3a4.pth` is a pretrained ConvNets for pytorch ResNeXt used by BGNN-trait-segmentation.
19
  See [github.com/Cadene/pretrained-models.pytorch#resnext](https://github.com/Cadene/pretrained-models.pytorch#resnext) for documentation about the source.
20
 
21
- The segmentation model was first trained on ImageNet ([Deng et al., 2009](10.1109/CVPR.2009.5206848)), and then the model was fine-tuned on a specific set of image data relevant to the domain: [Illinois Natural History Survey Fish Collection](https://fish.inhs.illinois.edu/) (INHS Fish).
22
  The Feature Pyramid Network (FPN) architecture was used for fine-tuning, since it is a CNN-based architecture designed to handle multi-scale feature maps (Lin et al., 2017: [IEEE](10.1109/CVPR.2017.106), [arXiv](arXiv:1612.03144)).
23
- The FPN uses SE-ResNeXt as the base network (Hu et al., 2018: [IEEE](10.1109/CVPR.2018.00745), [arXiv](arXiv:1709.01507)).
24
 
25
 
26
  ### Model Description
@@ -103,7 +103,7 @@ See instructions for use [here](https://github.com/hdr-bgnn/BGNN-trait-segmentat
103
 
104
  ## Training Details
105
 
106
- The image data were annotated using [SlicerMorph](https://slicermorph.github.io/) (Rolfe et al., 2021) by collaborators W. Dahdul and K. Diamond.
107
 
108
  ### Training Data
109
 
@@ -130,7 +130,7 @@ During the fine-tuning procedure, the encoder of the pre-trained model was froze
130
  We only tuned the decoder weights of our segmentation model during this fine-tuning procedure.
131
 
132
  We then trained the prepared model for 120 epochs, updating the weights using dice loss as a measure of similarity between the predicted and ground-truth segmentation.
133
- The Adam optimizer (Kingma & Ba, 2014) with a small learning rate (1e-4) was used to update the model weights.
134
 
135
 
136
  #### Preprocessing [optional]
 
18
  `se_resnext50_32x4d-a260b3a4.pth` is a pretrained ConvNets for pytorch ResNeXt used by BGNN-trait-segmentation.
19
  See [github.com/Cadene/pretrained-models.pytorch#resnext](https://github.com/Cadene/pretrained-models.pytorch#resnext) for documentation about the source.
20
 
21
+ The segmentation model was first trained on ImageNet ([Deng et al., 2009](https://doi.org/10.1109/CVPR.2009.5206848)), and then the model was fine-tuned on a specific set of image data relevant to the domain: [Illinois Natural History Survey Fish Collection](https://fish.inhs.illinois.edu/) (INHS Fish).
22
  The Feature Pyramid Network (FPN) architecture was used for fine-tuning, since it is a CNN-based architecture designed to handle multi-scale feature maps (Lin et al., 2017: [IEEE](10.1109/CVPR.2017.106), [arXiv](arXiv:1612.03144)).
23
+ The FPN uses SE-ResNeXt as the base network (Hu et al., 2018: [IEEE](https://doi.org/10.1109/CVPR.2018.00745), [arXiv](arXiv:1709.01507)).
24
 
25
 
26
  ### Model Description
 
103
 
104
  ## Training Details
105
 
106
+ The image data were annotated using [SlicerMorph](https://slicermorph.github.io/) ([Rolfe et al., 2021](https://doi.org/10.1111/2041-210X.13669)) by collaborators W. Dahdul and K. Diamond.
107
 
108
  ### Training Data
109
 
 
130
  We only tuned the decoder weights of our segmentation model during this fine-tuning procedure.
131
 
132
  We then trained the prepared model for 120 epochs, updating the weights using dice loss as a measure of similarity between the predicted and ground-truth segmentation.
133
+ The Adam optimizer ([Kingma & Ba, 2014](https://doi.org/10.48550/arXiv.1412.6980)) with a small learning rate (1e-4) was used to update the model weights.
134
 
135
 
136
  #### Preprocessing [optional]