gpt-oss-20b-lora-finetuned_fp4_step_20
This is a merged model combining GPT-OSS-20B with a fine-tuned LoRA adapter.
Model Details
- Base Model: openai/gpt-oss-20b
- LoRA Checkpoint: checkpoint-20
- Model Type: Causal Language Model
- Model Size: ~20B parameters
- Tensor Type: bfloat16
LoRA Configuration
- Rank (r): 8
- Alpha: 16
- Target Modules: k_proj, v_proj, o_proj, q_proj
- Special MLP Expert Layers: Layers 7, 15, 23
Quick Start
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained(
"ChickenMcSwag/gpt-oss-20b-lora-finetuned_fp4_step_20",
torch_dtype="auto",
device_map="auto",
trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained("ChickenMcSwag/gpt-oss-20b-lora-finetuned_fp4_step_20")
# Generate text
prompt = "The future of AI is"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
**inputs,
max_length=100,
temperature=0.7,
do_sample=True,
top_p=0.95
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Hardware Requirements
- Minimum VRAM: ~40GB for inference
- Recommended: 2x A100 80GB or equivalent
License
This model follows the original GPT-OSS-20B license. Please refer to the base model's license and usage policy.
Citation
If you use this model, please cite the original GPT-OSS-20B model.
- Downloads last month
- 3
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for ChickenMcSwag/gpt-oss-20b-lora-finetuned_fp4_step_20
Base model
openai/gpt-oss-20b