metadata
license: other
license_name: tongyi-qianwen
license_link: https://huggingface.co/Qwen/Qwen1.5-7B-Chat/blob/main/LICENSE
language:
- en
pipeline_tag: text-generation
tags:
- code
datasets:
- reciprocate/dpo_ultra-capybara-code_filtered-best
Coder1.8-ORPO-TEST
Model Description
Test model for ORPO finetune method, trained on ~20k code examples for 1 epoch on 2 x A40 cards with 4-bit QLora (lora rank=lora alpha=16).
Disclaimer
This is a test model and may generate incorrect responses. Use at your own risk.
Train Details
- Base: Qwen1.5-1.8B
- Training Data: ~20k code examples
- Epochs: 1
- Method: ORPO
- Hardware: 2 x A40
- Quantization: 4-bit QLora
- Lora Rank/Alpha: 16
Limitations
Limited training data and quantization may impact performance.
Join the Discussion
Have questions or feedback? Join our Discord server Here.