LoRA's
Collection
4 items
•
Updated
•
1
Hey! These LoRA adapters are trained using different reasoning datasets that utilize Thought and Solution for reasoning responses.
I hope these help jumpstart your project! These adapters have been trained on an A800 GPU and should provide a solid base for fine-tuning or merging.
Everything on my page is left public for Open Source use.
Here are the links to the available adapters as of January 30, 2025:
These adapters can be loaded and used with peft
and transformers
. Here’s a quick example:
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
base_model = "microsoft/Phi-4"
lora_adapter = "Quazim0t0/Phi4.Turn.R1Distill-Lora1"
tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, lora_adapter)
model.eval()