66 lines
2.8 KiB
Markdown
66 lines
2.8 KiB
Markdown
|
|
---
|
|||
|
|
license: apache-2.0
|
|||
|
|
license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE
|
|||
|
|
pipeline_tag: text-generation
|
|||
|
|
base_model:
|
|||
|
|
- huihui-ai/Qwen3-0.6B-abliterated
|
|||
|
|
tags:
|
|||
|
|
- chat
|
|||
|
|
- abliterated
|
|||
|
|
- uncensored
|
|||
|
|
extra_gated_prompt: >-
|
|||
|
|
**Usage Warnings**
|
|||
|
|
|
|||
|
|
|
|||
|
|
“**Risk of Sensitive or Controversial Outputs**“: This model’s safety
|
|||
|
|
filtering has been significantly reduced, potentially generating sensitive,
|
|||
|
|
controversial, or inappropriate content. Users should exercise caution and
|
|||
|
|
rigorously review generated outputs.
|
|||
|
|
|
|||
|
|
“**Not Suitable for All Audiences**:“ Due to limited content filtering, the
|
|||
|
|
model’s outputs may be inappropriate for public settings, underage users, or
|
|||
|
|
applications requiring high security.
|
|||
|
|
|
|||
|
|
“**Legal and Ethical Responsibilities**“: Users must ensure their usage
|
|||
|
|
complies with local laws and ethical standards. Generated content may carry
|
|||
|
|
legal or ethical risks, and users are solely responsible for any consequences.
|
|||
|
|
|
|||
|
|
“**Research and Experimental Use**“: It is recommended to use this model for
|
|||
|
|
research, testing, or controlled environments, avoiding direct use in
|
|||
|
|
production or public-facing commercial applications.
|
|||
|
|
|
|||
|
|
“**Monitoring and Review Recommendations**“: Users are strongly advised to
|
|||
|
|
monitor model outputs in real-time and conduct manual reviews when necessary
|
|||
|
|
to prevent the dissemination of inappropriate content.
|
|||
|
|
|
|||
|
|
“**No Default Safety Guarantees**“: Unlike standard models, this model has not
|
|||
|
|
undergone rigorous safety optimization. huihui.ai bears no responsibility for
|
|||
|
|
any consequences arising from its use.
|
|||
|
|
---
|
|||
|
|
|
|||
|
|
# Melvin56/Qwen3-0.6B-abliterated-GGUF
|
|||
|
|
|
|||
|
|
Original Model : [huihui-ai/Qwen3-0.6B-abliterated](https://huggingface.co/huihui-ai/Qwen3-0.6B-abliterated)
|
|||
|
|
|
|||
|
|
Llama.cpp build: 0208355 (5342)
|
|||
|
|
|
|||
|
|
I used imatrix to create all these quants using this [Dataset](https://gist.github.com/tristandruyen/9e207a95c7d75ddf37525d353e00659c/#file-calibration_data_v5_rc-txt).
|
|||
|
|
|
|||
|
|
---
|
|||
|
|
|
|||
|
|
| | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
|
|||
|
|
| :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
|
|||
|
|
| K-quants | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ 🐢5 | ✅ 🐢5 | ❌ |
|
|||
|
|
| I-quants | ✅ 🐢4 | ✅ 🐢4 | ✅ 🐢4 | ✅ | ✅ | Partial¹ | ❌ | ❌ | ❌ |
|
|||
|
|
```
|
|||
|
|
✅: feature works
|
|||
|
|
🚫: feature does not work
|
|||
|
|
❓: unknown, please contribute if you can test it youself
|
|||
|
|
🐢: feature is slow
|
|||
|
|
¹: IQ3_S and IQ1_S, see #5886
|
|||
|
|
²: Only with -ngl 0
|
|||
|
|
³: Inference is 50% slower
|
|||
|
|
⁴: Slower than K-quants of comparable size
|
|||
|
|
⁵: Slower than cuBLAS/rocBLAS on similar cards
|
|||
|
|
⁶: Only q8_0 and iq4_nl
|
|||
|
|
```
|