|
--- |
|
license: mit |
|
datasets: |
|
- agicorp/python_code_instructions_18k_alpaca |
|
language: |
|
- en |
|
base_model: |
|
- openai-community/gpt2 |
|
pipeline_tag: text-generation |
|
library_name: transformers |
|
tags: |
|
- code |
|
--- |
|
# Model Card |
|
|
|
GPT2Coder is a language model that uses openAI's GPT2 model architecture, |
|
the model was pre-trained on multiple code data focused on python and languages |
|
such as Spanish and English. |
|
|
|
- It is a pre-trained model in a medium amount of code, so it is not recommended to use it like this, but it is functional and serves uses such as fine tuning and other tasks. |
|
|
|
## Model Details |
|
|
|
|
|
- **Developed by:** BueormAI |
|
- **Shared by:** BueormLLC |
|
- **Model type:** Transformer |
|
- **Language(s) (NLP):** English (en), Spanish (es) |
|
- **License:** MiT |
|
- **Finetuned from model:** GPT2 Architecture |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
The model can generate unexpected code and output, in addition to offensive texts and non-functional code. |
|
|
|
|
|
### Recommendations |
|
|
|
We recommend using the model with caution and handling its outputs with discretion as they may turn out to be non-functional outputs and harmful and dangerous code. |
|
|
|
## Training Details |
|
|
|
### Training Hyperparameters |
|
|
|
- **Training regime:** fp16 mixed precision |
|
- **Max_lenght:** 1024 tokens |
|
- **pretrain epochs:** 1 epochs |
|
- **finetuning epochs:** 2 epochs |
|
|
|
## Environmental Impact |
|
|
|
- **Hardware Type:** GPU P100 |
|
- **Hours used:** 18 hours |
|
- **Cloud Provider:** Kaggle |
|
|
|
|
|
# By Bueorm |
|
Thanks to all the people who download and support our projects |
|
and manage a vision towards the future with AI, we hope you will support |
|
us to continue advancing and launching more followed models. |
|
|
|
- [Paypal Donations](https://paypal.me/bueorm) |
|
- [Patreon Subscription](https://patreon.com/bueorm) |