Model Card for Model ID
This LoRA adapter was extracted from mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated and uses meta-llama/Llama-3.1-70B-Instruct as a base.
Model Details
The model was extracted by running mlabonne/harmful_behaviors and the user prompts (but not assistant responses or system messages) from Guilherme34/uncensor through the original abliterated model to generate a dataset of prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl using FSDP. Since the original abliterated model isn't perfect at avoiding refusals, the dataset was cleaned to remove the few refusals generated prior to training.
Model Description
- Developed by: @reissbaker
- Funded by: Synthetic Lab
- License: Apache 2.0
- Finetuned from model: Llama 3.1 70B Instruct
How to Get Started with the Model
Run the model with one click on glhf.chat.
Training Hyperparameters
- BF16 mixed-precision
- 4e-4 LR
- Linear LR schedule
- Fused AdamW optimizer
- Downloads last month
- 607
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model’s pipeline type.
Model tree for reissbaker/llama-3.1-70b-abliterated-lora
Base model
meta-llama/Llama-3.1-70B
Finetuned
meta-llama/Llama-3.1-70B-Instruct