TissueGPT: Fine-Tuned BioGPT for Biomedical Text Generation

Model Description

TissueGPT is a fine-tuned version of BioGPT, specifically tailored for biomedical text generation tasks. By leveraging a dataset of biomedical research articles (titles, abstracts, and full texts), TissueGPT is designed to perform tasks such as:

  • Summarizing biomedical literature
  • Generating coherent biomedical text
  • Assisting with scientific writing in life sciences
  • Supporting research in tissue engineering, extracellular matrix (ECM) analysis, and related fields

Training Details

First Round of Training

The initial model was fine-tuned for 3 epochs, focusing on general adaptation to the biomedical dataset.

Hyperparameters

  • Learning Rate: 5e-5
  • Batch Size: 8
  • Warmup Steps: 500
  • Precision: Mixed precision (fp16)
  • Weight Decay: 0.01
  • Number of Epochs: 3
  • Save Checkpoints: Every 10,000 steps, keeping the last 3 checkpoints

Training and Validation Metrics

Epoch Training Loss Validation Loss Perplexity
1 2.4752 2.4286 11.34
2 2.3680 2.3708 10.70
3 2.2954 2.3410 10.39

Second Round of Training

To further improve performance, the model was fine-tuned for 2 additional epochs with adjusted hyperparameters.

Adjusted Hyperparameters

  • Learning Rate: 3e-5 (reduced for finer updates)
  • Batch Size: 64 (to utilize the GPU’s full memory)
  • Precision: bf16 (optimized for NVIDIA A100)
  • Save Checkpoints: Every 20,000 steps

Training and Validation Metrics

Epoch Training Loss Validation Loss Perplexity
4 2.2396 2.2395 9.43
5 2.2328 2.2328 9.32

Hardware Used

  • GPU: NVIDIA A100 80GB
  • Framework: PyTorch with Hugging Face Transformers library

Evaluation Metrics

Perplexity

Perplexity is a key metric for evaluating language models, measuring how well the model predicts sequences of text. Lower perplexity indicates better predictive performance.

  • First Round of Training: Final perplexity = 10.39
  • Second Round of Training: Final perplexity = 9.32

A lower perplexity indicates that the model generates more fluent and coherent text.

Gradient Norms

  • Tracked gradient stability during training.
  • Observed Range: 1.05–1.32, indicating stable training.

Validation Loss

  • Decreasing validation loss across both rounds suggests effective generalization to unseen data.

Model Comparison

Metric First Round Second Round
Final Validation Loss 2.3410 2.2328
Final Perplexity 10.39 9.32

Key Insights:

  • Additional training epochs led to improved generalization and better predictive performance.
  • Perplexity improved by approximately 10% in the second round, demonstrating enhanced text fluency and coherence.

How to Use the Model

Install Dependencies

Ensure you have transformers and torch installed:

pip install transformers torch

Load the Model

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "Saeed/TissueGPT"  # Replace with the uploaded repo name
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

input_text = "The extracellular matrix plays a critical role in tissue engineering because"
inputs = tokenizer(input_text, return_tensors="pt")

output = model.generate(**inputs, max_length=50)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Intended Use

  • Biomedical text generation and summarization
  • Assisting researchers, scientists, and medical professionals
  • Automated scientific writing in domains like tissue engineering, and scaffold fabrication.

Limitations

  • The model is fine-tuned on biomedical literature and may not generalize well to non-biomedical domains.
  • Outputs should always be validated by experts for accuracy, especially in clinical or research-critical contexts.

Ethical Considerations

  • This model is intended for use in biomedical research and not for clinical diagnosis or patient care.
  • It may generate plausible-sounding but factually incorrect outputs (hallucinations). Always verify generated content.

Citation

If you use TissueGPT, please cite the following:

The citation details will be provided shortly.

License

Licensed under the CC BY 4.0 License.

Contact

For questions, issues, or collaboration opportunities, feel free to reach out at:

Downloads last month
1
Safetensors
Model size
347M params
Tensor type
F32
·
Inference Examples
Unable to determine this model's library. Check the docs .