models: | |
- model: CultriX/Qwen2.5-14B-Wernickev3 | |
parameters: | |
weight: 0.38 # Slight reduction to balance with FinalMerge's generalist capabilities. | |
density: 0.65 # Retain significant parameters for stability and strong task performance. | |
- model: CultriX/Qwen2.5-14B-FinalMerge | |
parameters: | |
weight: 0.32 # Slight increase to ensure its generalist capabilities are fully utilized. | |
density: 0.60 # Balanced density for comprehensive task coverage. | |
- model: CultriX/Qwen2.5-14B-Emergedv3 | |
parameters: | |
weight: 0.20 # Retains focused contribution to specific task optimizations. | |
density: 0.55 # Moderate density ensures efficient parameter usage. | |
- model: qingy2019/Qwen2.5-Math-14B-Instruct | |
parameters: | |
weight: 0.10 # Consistent with its specialist focus, balancing lower weight with higher density. | |
density: 0.70 # High density ensures retention of advanced reasoning and MATH-related parameters. | |
merge_method: dare_ties | |
base_model: CultriX/SeQwence-14Bv1 | |
parameters: | |
normalize: true # Ensures all models are scaled to compatible parameter ranges. | |
int8_mask: true # Optimizes memory and computational efficiency without accuracy loss. | |
dtype: bfloat16 # Provides better memory efficiency and numerical stability. | |
adaptive_merge_parameters: | |
task_weights: | |
tinyArc: 1.3 # Slight reduction to balance with generalist contributions. | |
tinyHellaswag: 1.3 # Maintains strong performance in contextual reasoning. | |
tinyMMLU: 1.2 # Balanced focus for domain-specific knowledge. | |
tinyTruthfulQA: 1.2 # Adjusted to ensure fair contribution without over-prioritization. | |
tinyTruthfulQA_mc1: 1.1 # Maintains a moderate priority to balance with other tiny benchmarks. | |
tinyWinogrande: 1.2 # Strong contextual reasoning support from generalist models. | |
IFEval: 1.5 # High weight for general instruction-following capabilities. | |
BBH: 1.5 # Prioritizes complex reasoning and multi-step problem-solving tasks. | |
MATH: 1.55 # Slight reduction to balance MATH with other advanced reasoning benchmarks. | |
GPQA: 1.4 # Balanced to reflect contributions from both generalist and specialist models. | |
MUSR: 1.4 # Increased slightly to strengthen multi-step reasoning. | |
MMLU-PRO: 1.3 # Maintains general task performance across multitask domain knowledge. | |
smoothing_factor: 0.18 # Slightly increased for smoother blending across task boundaries. | |
gradient_clipping: 0.88 # Tightened slightly for stability, preventing parameter over-contribution. | |