license, license_link, pipeline_tag, base_model, tags, extra_gated_prompt
license license_link pipeline_tag base_model tags extra_gated_prompt
apache-2.0 https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE text-generation
huihui-ai/Qwen3-0.6B-abliterated
chat
abliterated
uncensored
**Usage Warnings** “**Risk of Sensitive or Controversial Outputs**“: This models safety filtering has been significantly reduced, potentially generating sensitive, controversial, or inappropriate content. Users should exercise caution and rigorously review generated outputs. “**Not Suitable for All Audiences**:“ Due to limited content filtering, the models outputs may be inappropriate for public settings, underage users, or applications requiring high security. “**Legal and Ethical Responsibilities**“: Users must ensure their usage complies with local laws and ethical standards. Generated content may carry legal or ethical risks, and users are solely responsible for any consequences. “**Research and Experimental Use**“: It is recommended to use this model for research, testing, or controlled environments, avoiding direct use in production or public-facing commercial applications. “**Monitoring and Review Recommendations**“: Users are strongly advised to monitor model outputs in real-time and conduct manual reviews when necessary to prevent the dissemination of inappropriate content. “**No Default Safety Guarantees**“: Unlike standard models, this model has not undergone rigorous safety optimization. huihui.ai bears no responsibility for any consequences arising from its use.

Melvin56/Qwen3-0.6B-abliterated-GGUF

Original Model : huihui-ai/Qwen3-0.6B-abliterated

Llama.cpp build: 0208355 (5342)

I used imatrix to create all these quants using this Dataset.


CPU (AVX2) CPU (ARM NEON) Metal cuBLAS rocBLAS SYCL CLBlast Vulkan Kompute
K-quants 🐢5 🐢5
I-quants 🐢4 🐢4 🐢4 Partial¹
✅: feature works
🚫: feature does not work
❓: unknown, please contribute if you can test it youself
🐢: feature is slow
¹: IQ3_S and IQ1_S, see #5886
²: Only with -ngl 0
³: Inference is 50% slower
⁴: Slower than K-quants of comparable size
⁵: Slower than cuBLAS/rocBLAS on similar cards
⁶: Only q8_0 and iq4_nl
Description
Model synced from source: Melvin56/Qwen3-0.6B-abliterated-GGUF
Readme 26 KiB