You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Qwen2.5-7b-instruct-sft

This is a fine-tuned version of the Qwen2.5-7B-Instruct model for instruction-following tasks. It was fine-tuned using the SFTTrainer from the trl library on the OpenAssistant Guanaco dataset.

Model Details

Base Model

  • Model: Qwen2.5-7B-Instruct
  • Architecture: Transformer-based causal language model
  • License: Apache 2.0

Fine-Tuning Details

  • Dataset: OpenAssistant Guanaco
  • Training Epochs: 1
  • Batch Size: 2
  • Gradient Accumulation Steps: 16
  • Learning Rate: 1e-5
  • Optimizer: Paged AdamW 8-bit
  • Mixed Precision: fp16 (if bf16 is not supported) or bf16
  • Max Sequence Length: 512 tokens

Training Hardware

  • GPU: NVIDIA A100 (or your specific GPU)
  • Training Time: X hours (optional)

Usage

You can use this model with the Hugging Face transformers library:

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the model and tokenizer
model = AutoModelForCausalLM.from_pretrained("Ameer-linnk/qwen2.5-7b-instruct-sft")
tokenizer = AutoTokenizer.from_pretrained("Ameer-linnk/qwen2.5-7b-instruct-sft")

# Prepare input
input_text = "What is the capital of France?"
inputs = tokenizer(input_text, return_tensors="pt")

# Generate output
outputs = model.generate(**inputs)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response) 
Example Output
Input:
What is the capital of France?

Output:
The capital of France is Paris.
Downloads last month
4
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference API
Unable to determine this model's library. Check the docs .