66 lines
2.8 KiB
Markdown
66 lines
2.8 KiB
Markdown
---
|
||
license: apache-2.0
|
||
license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE
|
||
pipeline_tag: text-generation
|
||
base_model:
|
||
- huihui-ai/Qwen3-0.6B-abliterated
|
||
tags:
|
||
- chat
|
||
- abliterated
|
||
- uncensored
|
||
extra_gated_prompt: >-
|
||
**Usage Warnings**
|
||
|
||
|
||
“**Risk of Sensitive or Controversial Outputs**“: This model’s safety
|
||
filtering has been significantly reduced, potentially generating sensitive,
|
||
controversial, or inappropriate content. Users should exercise caution and
|
||
rigorously review generated outputs.
|
||
|
||
“**Not Suitable for All Audiences**:“ Due to limited content filtering, the
|
||
model’s outputs may be inappropriate for public settings, underage users, or
|
||
applications requiring high security.
|
||
|
||
“**Legal and Ethical Responsibilities**“: Users must ensure their usage
|
||
complies with local laws and ethical standards. Generated content may carry
|
||
legal or ethical risks, and users are solely responsible for any consequences.
|
||
|
||
“**Research and Experimental Use**“: It is recommended to use this model for
|
||
research, testing, or controlled environments, avoiding direct use in
|
||
production or public-facing commercial applications.
|
||
|
||
“**Monitoring and Review Recommendations**“: Users are strongly advised to
|
||
monitor model outputs in real-time and conduct manual reviews when necessary
|
||
to prevent the dissemination of inappropriate content.
|
||
|
||
“**No Default Safety Guarantees**“: Unlike standard models, this model has not
|
||
undergone rigorous safety optimization. huihui.ai bears no responsibility for
|
||
any consequences arising from its use.
|
||
---
|
||
|
||
# Melvin56/Qwen3-0.6B-abliterated-GGUF
|
||
|
||
Original Model : [huihui-ai/Qwen3-0.6B-abliterated](https://huggingface.co/huihui-ai/Qwen3-0.6B-abliterated)
|
||
|
||
Llama.cpp build: 0208355 (5342)
|
||
|
||
I used imatrix to create all these quants using this [Dataset](https://gist.github.com/tristandruyen/9e207a95c7d75ddf37525d353e00659c/#file-calibration_data_v5_rc-txt).
|
||
|
||
---
|
||
|
||
| | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
|
||
| :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
|
||
| K-quants | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ 🐢5 | ✅ 🐢5 | ❌ |
|
||
| I-quants | ✅ 🐢4 | ✅ 🐢4 | ✅ 🐢4 | ✅ | ✅ | Partial¹ | ❌ | ❌ | ❌ |
|
||
```
|
||
✅: feature works
|
||
🚫: feature does not work
|
||
❓: unknown, please contribute if you can test it youself
|
||
🐢: feature is slow
|
||
¹: IQ3_S and IQ1_S, see #5886
|
||
²: Only with -ngl 0
|
||
³: Inference is 50% slower
|
||
⁴: Slower than K-quants of comparable size
|
||
⁵: Slower than cuBLAS/rocBLAS on similar cards
|
||
⁶: Only q8_0 and iq4_nl
|
||
``` |