IMPORTANT: This model has an upgraded Jinja template which repairs issues with org model (repeats, long thinking, loops) and upgrades/repairs to tool handling.

Qwen3.5-21B-Claude-4.6-Opus-Thinking-EXP2

Experimental "Honey I Shrunk the Model" using Qwen 3.5 27B dense as a base, followed by training on Claude 4.6 Opus (small) dataset.

This process removed 16 layers from the model, with training via unsloth to re-stablize/re-train the model.

EXP2 used a different method to "shrink the model".

Not everything may work.

Please report any issues via the community tab.

Thank you.

More experiments in progress.

Downloads last month
17
Safetensors
Model size
21B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DavidAU/Qwen3.5-21B-Claude-4.6-Opus-Thinking-EXP2

Base model

Qwen/Qwen3.5-27B
Finetuned
(91)
this model
Quantizations
2 models

Collections including DavidAU/Qwen3.5-21B-Claude-4.6-Opus-Thinking-EXP2