🎛️Fine-tuning LLMs
Collection
Collection of fine-tuned LLMs.
•
29 items
•
Updated
•
1
This model is a fine-tuned version of openai/gpt-oss-20b on the kingabzpro/dermatology-qa-firecrawl-dataset dataset. It has been trained using TRL.
from transformers import pipeline
question = "How does the source suggest clinicians approach the diagnosis of rosacea?"
# Load pipeline
generator = pipeline(
"text-generation",
model="kingabzpro/gpt-oss-20b-dermatology-qa",
device="cuda" # or device=0
)
# Run inference (passing in chat-style format)
output = generator(
[{"role": "user", "content": question}],
max_new_tokens=200,
return_full_text=False
)[0]
print(output["generated_text"])
# The source says that clinicians should use a combination of clinical signs and symptoms when diagnosing rosacea, rather than relying on a single feature.
This model was trained with SFT.
Cite TRL as:
@misc{vonwerra2022trl,
title = {{TRL: Transformer Reinforcement Learning}},
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
year = 2020,
journal = {GitHub repository},
publisher = {GitHub},
howpublished = {\url{https://github.com/huggingface/trl}}
}
Base model
openai/gpt-oss-20b