Qwen2.5-14B-FinalMergev2 / mergekit_config.yml
CultriX's picture
Upload folder using huggingface_hub
3dcf22b verified
raw
history blame
2.67 kB
models:
- model: CultriX/Qwen2.5-14B-Wernickev3
parameters:
weight: 0.38 # Slight reduction to balance with FinalMerge's generalist capabilities.
density: 0.65 # Retain significant parameters for stability and strong task performance.
- model: CultriX/Qwen2.5-14B-FinalMerge
parameters:
weight: 0.32 # Slight increase to ensure its generalist capabilities are fully utilized.
density: 0.60 # Balanced density for comprehensive task coverage.
- model: CultriX/Qwen2.5-14B-Emergedv3
parameters:
weight: 0.20 # Retains focused contribution to specific task optimizations.
density: 0.55 # Moderate density ensures efficient parameter usage.
- model: qingy2019/Qwen2.5-Math-14B-Instruct
parameters:
weight: 0.10 # Consistent with its specialist focus, balancing lower weight with higher density.
density: 0.70 # High density ensures retention of advanced reasoning and MATH-related parameters.
merge_method: dare_ties
base_model: CultriX/SeQwence-14Bv1
parameters:
normalize: true # Ensures all models are scaled to compatible parameter ranges.
int8_mask: true # Optimizes memory and computational efficiency without accuracy loss.
dtype: bfloat16 # Provides better memory efficiency and numerical stability.
adaptive_merge_parameters:
task_weights:
tinyArc: 1.3 # Slight reduction to balance with generalist contributions.
tinyHellaswag: 1.3 # Maintains strong performance in contextual reasoning.
tinyMMLU: 1.2 # Balanced focus for domain-specific knowledge.
tinyTruthfulQA: 1.2 # Adjusted to ensure fair contribution without over-prioritization.
tinyTruthfulQA_mc1: 1.1 # Maintains a moderate priority to balance with other tiny benchmarks.
tinyWinogrande: 1.2 # Strong contextual reasoning support from generalist models.
IFEval: 1.5 # High weight for general instruction-following capabilities.
BBH: 1.5 # Prioritizes complex reasoning and multi-step problem-solving tasks.
MATH: 1.55 # Slight reduction to balance MATH with other advanced reasoning benchmarks.
GPQA: 1.4 # Balanced to reflect contributions from both generalist and specialist models.
MUSR: 1.4 # Increased slightly to strengthen multi-step reasoning.
MMLU-PRO: 1.3 # Maintains general task performance across multitask domain knowledge.
smoothing_factor: 0.18 # Slightly increased for smoother blending across task boundaries.
gradient_clipping: 0.88 # Tightened slightly for stability, preventing parameter over-contribution.