Model Card for BioMistral-7B-Finetuned

Model Summary

BioMistral-7B-Finetuned is a biomedical language model adapted from the BioMistral-7B model. This fine-tuned model is tailored for biomedical question-answering tasks and optimized through LoRA (Low-Rank Adaptation) on a 4-bit quantized base. It is particularly useful for tasks that require understanding and generating biomedical text in English.


Model Details

Model Description

This model was fine-tuned for biomedical applications, primarily focusing on enhancing accuracy in question-answering tasks within this domain.

  • Base Model: BioMistral-7B
  • License: apache-2.0
  • Fine-tuned for Task: Biomedical Q&A, text generation
  • Quantization: 4-bit precision with BitsAndBytes for efficient deployment

Uses

Direct Use

The model is suitable for biomedical question-answering and other related language generation tasks.

Out-of-Scope Use

Not recommended for general-purpose NLP tasks outside the biomedical domain or for clinical decision-making.


How to Get Started with the Model

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("BeastGokul/BioMistral-7B-Finetuned")
model = AutoModelForCausalLM.from_pretrained("BeastGokul/BioMistral-7B-Finetuned")

# Example usage
input_text = "What are the symptoms of diabetes?"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

Training Procedure

The model was fine-tuned using the LoRA (Low-Rank Adaptation) method, with a configuration set for biomedical question-answering.

Training Hyperparameters Precision: 4-bit quantization with BitsAndBytes Learning Rate: 2e-5 Batch Size: Effective batch size of 16 (4 per device, gradient accumulation steps of 4) Number of Epochs: 3

Framework versions

PEFT 0.13.2

Downloads last month
0
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for BeastGokul/Bio-Mistral-7B-finetuned

Adapter
(14)
this model