🧠 Qwen 3.5 Abliterated for ComfyUI (MXFP8 & NVFP4)

Welcome! This repository provides ComfyUI-ready, abliterated versions of Qwen 3.5, optimized for local AI workflows, assistants, and multimodal use inside ComfyUI.

📦 Quantized Model Files

Model Precision & Notes Approx Size
Heretical‑Qwen3.5‑9B‑fp8.safetensors FP8 quantized ablated Qwen‑3.5 9B 11.9 GB
qwen3.5_9b_abliterated_nvfp4.safetensors NVFP4 quantized ablated Qwen‑3.5 9B 8.36 GB
Qwen3.5‑4B‑heretic‑fp8.safetensors FP8 quantized ablated Qwen‑3.5 4B 5.51 GB
qwen3.5_4b_nvfp4.safetensors NVFP4 quantized Qwen‑3.5 4B 3.54 GB
qwen3.5_4b_claude46opus_abliterated_mxfp8mixedfp8.safetensors Mixed FP8 ablated (Claude‑4.6+Opus style) 5.91 GB
qwen3.5_4b_claude46opus_abliterated_nvfp4.safetensors NVFP4 variant of the above 3.54 GB

Example of simple workflow using Qwen 3.5 as an image assistant⬇️⬇️⬇️⬇️⬇️

Example of simple workflow plus image captioning⬇️⬇️⬇️⬇️⬇️

🚀 Overview These models are based on Qwen 3.5, a powerful multilingual LLM family designed for reasoning, coding, and general AI tasks.

This release includes:

🔓 Abliterated variants (uncensored / no refusal behavior) ⚡ MXFP8 & NVFP4 quantizations 🧩 Native ComfyUI compatibility

Abliteration removes built-in refusal behavior without retraining, preserving most of the model’s original capabilities while enabling unrestricted outputs. ✨ Features 🧠 Full AI assistant inside ComfyUI 🖼️ Image understanding (multimodal support) ⚡ Fast inference with low VRAM usage 🔌 Plug-and-play with standard ComfyUI nodes

These models can: Answer questions Generate prompts Assist workflows Analyze images directly inside ComfyUI pipelines ⚙️ Quantization Types 🔹 MXFP8 Balanced performance and quality Works on a wide range of GPUs Ideal default option 🔹 NVFP4 Ultra-low precision (4-bit) Massive VRAM reduction and speed gains Best suited for newer NVIDIA architectures Designed for efficient deployment of LLMs with minimal memory footprint 🧩 ComfyUI Integration

✅ These models load directly using:

CLIP Loader (standard node)

No special loaders required.

📦 Installation Update ComfyUI to the latest version Download the model file (MXFP8 or NVFP4)

Place it in:

ComfyUI/models/clip/ Load it using CLIP Loader 🎬 Workflow

👉 A workflow is provided in this repo to help you get started.

It demonstrates:

AI assistant usage Prompt generation Image interpretation

Highly recommended to download and test the workflow to understand full capabilities.

🧠 What is Abliteration?

Abliteration is a technique that:

Removes refusal/alignment layers Keeps original model intelligence intact Does not require retraining

Result: 👉 More freedom 👉 Same core performance

💡 Use Cases 🎥 Prompt generation for video models (LTX, WAN, etc.) 🧩 ComfyUI automation assistant 🖼️ Image captioning & interpretation ✍️ Creative writing / uncensored outputs 🧠 Local AI copilots ⚠️ Notes Abliterated models are community-created, not official releases Use responsibly depending on your application NVFP4 may require newer GPUs for best performance ❤️ Credits Base model: Qwen Team Quantization & conversion: community efforts ComfyUI integration: ongoing community development 🔥 Final Thoughts

If you want a fully local AI assistant embedded directly inside ComfyUI, this setup is one of the most powerful workflows available right now.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support