You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Qwen3.5-122B-A10B-abliterated

Unrestricted version of Qwen/Qwen3.5-122B-A10B, created with Abliterix โ€” automated LLM abliteration via orthogonalized steering and Bayesian optimization.

Highlights

Metric Value
Refusal rate 1/200 (0.5%)
KL divergence 0.0115
Optimization trials 25

The largest abliterated Qwen3.5 model. Only 1 out of 200 test prompts triggered a refusal โ€” a 0.5% refusal rate with near-zero model degradation.

How It Works

Abliterix removes safety-refusal behavior while preserving model capabilities:

  1. Refusal direction extraction โ€” 800 harmful + 800 benign prompts reveal per-layer refusal activation patterns
  2. Orthogonal projection โ€” isolates the refusal signal by projecting out components aligned with normal responses, reducing refusals by 67% vs. raw abliteration
  3. LoRA-based abliteration โ€” rank-1 modifications to attention and MLP weights, captured as lightweight adapters (not destructive edits)
  4. Bayesian optimization โ€” Optuna TPE searches kernel shape, fractional direction index, and per-component strength across 25 trials to find the Pareto-optimal balance of low refusals and low KL divergence

All Abliterix Models

Model Refusals KL Divergence Trials
Qwen3.5-122B-A10B-abliterated 1/200 (0.5%) 0.0115 25
Qwen3.5-35B-A3B-abliterated 3/200 (1.5%) 0.0035 50
Qwen3.5-27B-abliterated 3/200 (1.5%) 0.0051 35
Qwen3.5-9B-abliterated 2/200 (1%) 0.0105 50
Qwen3.5-4B-abliterated 3/200 (1.5%) 0.0065 50
Qwen3.5-0.8B-abliterated 0/200 (0%) 0.0087 100

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("wangzhang/Qwen3.5-122B-A10B-abliterated", torch_dtype="auto", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("wangzhang/Qwen3.5-122B-A10B-abliterated")

messages = [{"role": "user", "content": "Your question here"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True, enable_thinking=False)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Citation

@software{abliterix,
  author = {Wu, Wangzhang},
  title = {Abliterix: Automated LLM Abliteration},
  year = {2026},
  url = {https://github.com/wuwangzhang1216/abliterix}
}

Links


Built with Abliterix | PyPI

Downloads last month
1,618
Safetensors
Model size
122B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for wangzhang/Qwen3.5-122B-A10B-abliterix

Finetuned
(39)
this model
Quantizations
15 models

Collection including wangzhang/Qwen3.5-122B-A10B-abliterix