llama327kto / README.md
kevin009's picture
Update README.md
963b245 verified
metadata
license: apache-2.0
language:
  - en
base_model:
  - meta-llama/Llama-3.1-8B-instruct
pipeline_tag: text-generation
tags:
  - lora
  - adapter
  - writing
  - CoT

Model Details

  • Base Model: meta-llama/Llama-3.1-8B-instruct

Merger Configuration

Source Adapters

All source adapters share the following configuration:

  • Rank (r): 16
  • Alpha: 16
  • Target Modules:
    • q_proj (Query projection)
    • k_proj (Key projection)
    • v_proj (Value projection)
    • o_proj (Output projection)
    • up_proj (Upsampling projection)
    • down_proj (Downsampling projection)
    • gate_proj (Gate projection)
  • The order of loading adapters may affect the final result
  • Equal weights were chosen to maintain balanced influence from each adapter
  • The merged adapter maintains the same architecture and rank as the original adapters
  • While this adapter merges multiple fine-tunes, each component was developed as part of independent research efforts to explore and language model capabilities as part of R&D process.