SFT
Collection
Supervised fine-tuning
•
7 items
•
Updated
•
1
Bio-8B-it is an 8B parameter biomedical instruction-tuned language model built on top of Qwen 3-8B. The model was fine-tuned using Supervised Fine-Tuning (SFT) with QLoRA via the PEFT framework.
This model is optimized for biomedical and clinical NLP instruction-following tasks, including:
Base Model
Fine-Tuning Method
This model is intended for:
This model is not intended for:
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("khazarai/Bio-8B-it")
model = AutoModelForCausalLM.from_pretrained(
"khazarai/Bio-8B-it",
device_map={"": 0}
)
question = """
Describe how to properly perform a hand hygiene using an alcohol-based hand sanitizer.
"""
messages = [
{"role" : "user", "content" : question}
]
text = tokenizer.apply_chat_template(
messages,
tokenize = False,
add_generation_prompt = True,
enable_thinking = False,
)
from transformers import TextStreamer
_ = model.generate(
**tokenizer(text, return_tensors = "pt").to("cuda"),
max_new_tokens = 1400,
temperature = 0.7,
top_p = 0.8,
top_k = 20,
streamer = TextStreamer(tokenizer, skip_prompt = True),
)
Citation
If you use this model, please cite the original BioInstruct paper:
@article{Tran2024Bioinstruct,
author = {Tran, Hieu and Yang, Zhichao and Yao, Zonghai and Yu, Hong},
title = {BioInstruct: instruction tuning of large language models for biomedical natural language processing},
journal = {Journal of the American Medical Informatics Association},
year = {2024},
doi = {10.1093/jamia/ocae122}
}