--- license: apache-2.0 language: - en base_model: - meta-llama/Llama-3.1-8B-instruct pipeline_tag: text-generation tags: - lora - adapter - writing - CoT --- ## Model Details - Base Model: meta-llama/Llama-3.1-8B-instruct ## Merger Configuration ### Source Adapters All source adapters share the following configuration: - Rank (r): 16 - Alpha: 16 - Target Modules: - q_proj (Query projection) - k_proj (Key projection) - v_proj (Value projection) - o_proj (Output projection) - up_proj (Upsampling projection) - down_proj (Downsampling projection) - gate_proj (Gate projection) - The order of loading adapters may affect the final result - Equal weights were chosen to maintain balanced influence from each adapter - The merged adapter maintains the same architecture and rank as the original adapters - While this adapter merges multiple fine-tunes, each component was developed as part of independent research efforts to explore and language model capabilities as part of R&D process.