Phr00tyMix-v1-32B

Note: this model has been superseded by Phr00tyMix-v2

image/png

This is a merge of pre-trained language models created using mergekit.

The goal is to be smart, obedient, creative and coherent. This isn't 100% censored, but some simple prompting to disallow refusals seems to do the trick.

GGUFs can be found here

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using rombodawg/Rombos-LLM-V2.5-Qwen-32b as a base.

This base model was chosen as a smart, non-thinking foundation.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

merge_method: dare_ties
dtype: bfloat16
base_model: rombodawg/Rombos-LLM-V2.5-Qwen-32b
parameters:
  normalize_weights: true
models:
  - model: Delta-Vector/Hamanasu-QwQ-V1.5-Instruct
    parameters:
      weight: 0.3
      density: 1
  - model: zetasepic/Qwen2.5-32B-Instruct-abliterated-v2
    parameters:
      weight: 0.1
      density: 0.8
  - model: THU-KEG/LongWriter-Zero-32B
    parameters:
      weight: 0.1
      density: 0.8
  - model: Delta-Vector/Hamanasu-Magnum-QwQ-32B
    parameters:
      weight: 0.3
      density: 0.8
  - model: allura-org/Qwen2.5-32b-RP-Ink
    parameters:
      weight: 0.2
      density: 0.5
Downloads last month
12
Safetensors
Model size
32.8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Phr00t/Phr00tyMix-v1-32B