Model Card for Hatespeech and Offensive Classification Model

Model Details

Developed by: Marco Orasch

Model Type: Transformer-based Language Model

Language(s): English

Finetuned from model: llama-3.2-3B

Training Epochs: 1

Batch Size: 1

Learning Rate: 2e-4

Optimizer: AdamW

Model Description

This model is fine-tuned for the task of classifying hate speech and offensive language in English text. It leverages the llama-3.2-3B model architecture, refined through a supervised learning approach using a dataset specifically curated for identifying harmful and offensive content. The model can distinguish between hate speech, offensive language, and neutral content, making it suitable for content moderation and safety applications.

Training Procedure

Run Details: https://api.wandb.ai/links/marcoor-universit-t-klagenfurt/oxoxik67

Base Model: llama-3.2-3B

Fine-tuning Framework: PyTorch with Hugging Face Transformers

Tutorial Used: https://www.datacamp.com/tutorial/fine-tuning-llama-3-1

Jupyter Notebook: https://huggingface.co/marcoorasch/llama-3.2-3B-instruct-hatespeech-offensive-classification/blob/main/Hatespeech_Offensive_Classification_llama3.2-3B-instruct.ipynb

Intended Use

Content moderation on social media platforms

Automated filtering of offensive content in forums and chat applications

Assisting research in online safety and digital well-being

Downloads last month
23
Safetensors
Model size
3.21B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for marcoorasch/llama-3.2-3B-instruct-hatespeech-offensive-classification

Finetuned
(214)
this model

Dataset used to train marcoorasch/llama-3.2-3B-instruct-hatespeech-offensive-classification