--- base_model: - MaziyarPanahi/calme-2.4-rys-78B - Sakalti/ultiima-78B library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Linear DELLA](https://arxiv.org/abs/2406.11617) merge method using [Sakalti/ultiima-78B](https://huggingface.co/Sakalti/ultiima-78B) as a base. ### Models Merged The following models were included in the merge: * [MaziyarPanahi/calme-2.4-rys-78B](https://huggingface.co/MaziyarPanahi/calme-2.4-rys-78B) ### Configuration The following YAML configuration was used to produce this model: ```yaml merge_method: della_linear base_model: Sakalti/ultiima-78B dtype: float16 parameters: epsilon: 0.015 # Fine-grain scaling for precision. lambda: 1.6 # Strong emphasis on top-performing models. normalize: true # Stable parameter integration across models. adaptive_merge_parameters: task_weights: tinyArc: 1.75 # Logical reasoning. tinyHellaswag: 1.65 # Contextual predictions. tinyMMLU: 1.8 # Domain knowledge. tinyTruthfulQA: 2.0 # Prioritize truthful reasoning. tinyTruthfulQA_mc1: 1.85 tinyWinogrande: 1.9 # Advanced reasoning and predictions. IFEval: 2.1 # Instruction-following and multitasking. BBH: 1.9 # Complex reasoning. MATH: 2.3 # Mathematical reasoning. GPQA: 2.2 # Factual QA. MUSR: 2.0 # Multi-step reasoning. MMLU-PRO: 2.2 # Domain multitask performance. smoothing_factor: 0.1 # Smooth blending across benchmarks. models: - model: MaziyarPanahi/calme-2.4-rys-78B parameters: weight: 1 density: 1 - model: Sakalti/ultiima-78B parameters: weight: 1 density: 1 ```