初始化项目,由ModelHub XC社区提供模型
Model: Melvin56/Qwen3-0.6B-abliterated-GGUF Source: Original Platform
This commit is contained in:
66
README.md
Normal file
66
README.md
Normal file
@@ -0,0 +1,66 @@
|
||||
---
|
||||
license: apache-2.0
|
||||
license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE
|
||||
pipeline_tag: text-generation
|
||||
base_model:
|
||||
- huihui-ai/Qwen3-0.6B-abliterated
|
||||
tags:
|
||||
- chat
|
||||
- abliterated
|
||||
- uncensored
|
||||
extra_gated_prompt: >-
|
||||
**Usage Warnings**
|
||||
|
||||
|
||||
“**Risk of Sensitive or Controversial Outputs**“: This model’s safety
|
||||
filtering has been significantly reduced, potentially generating sensitive,
|
||||
controversial, or inappropriate content. Users should exercise caution and
|
||||
rigorously review generated outputs.
|
||||
|
||||
“**Not Suitable for All Audiences**:“ Due to limited content filtering, the
|
||||
model’s outputs may be inappropriate for public settings, underage users, or
|
||||
applications requiring high security.
|
||||
|
||||
“**Legal and Ethical Responsibilities**“: Users must ensure their usage
|
||||
complies with local laws and ethical standards. Generated content may carry
|
||||
legal or ethical risks, and users are solely responsible for any consequences.
|
||||
|
||||
“**Research and Experimental Use**“: It is recommended to use this model for
|
||||
research, testing, or controlled environments, avoiding direct use in
|
||||
production or public-facing commercial applications.
|
||||
|
||||
“**Monitoring and Review Recommendations**“: Users are strongly advised to
|
||||
monitor model outputs in real-time and conduct manual reviews when necessary
|
||||
to prevent the dissemination of inappropriate content.
|
||||
|
||||
“**No Default Safety Guarantees**“: Unlike standard models, this model has not
|
||||
undergone rigorous safety optimization. huihui.ai bears no responsibility for
|
||||
any consequences arising from its use.
|
||||
---
|
||||
|
||||
# Melvin56/Qwen3-0.6B-abliterated-GGUF
|
||||
|
||||
Original Model : [huihui-ai/Qwen3-0.6B-abliterated](https://huggingface.co/huihui-ai/Qwen3-0.6B-abliterated)
|
||||
|
||||
Llama.cpp build: 0208355 (5342)
|
||||
|
||||
I used imatrix to create all these quants using this [Dataset](https://gist.github.com/tristandruyen/9e207a95c7d75ddf37525d353e00659c/#file-calibration_data_v5_rc-txt).
|
||||
|
||||
---
|
||||
|
||||
| | CPU (AVX2) | CPU (ARM NEON) | Metal | cuBLAS | rocBLAS | SYCL | CLBlast | Vulkan | Kompute |
|
||||
| :------------ | :---------: | :------------: | :---: | :----: | :-----: | :---: | :------: | :----: | :------: |
|
||||
| K-quants | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ 🐢5 | ✅ 🐢5 | ❌ |
|
||||
| I-quants | ✅ 🐢4 | ✅ 🐢4 | ✅ 🐢4 | ✅ | ✅ | Partial¹ | ❌ | ❌ | ❌ |
|
||||
```
|
||||
✅: feature works
|
||||
🚫: feature does not work
|
||||
❓: unknown, please contribute if you can test it youself
|
||||
🐢: feature is slow
|
||||
¹: IQ3_S and IQ1_S, see #5886
|
||||
²: Only with -ngl 0
|
||||
³: Inference is 50% slower
|
||||
⁴: Slower than K-quants of comparable size
|
||||
⁵: Slower than cuBLAS/rocBLAS on similar cards
|
||||
⁶: Only q8_0 and iq4_nl
|
||||
```
|
||||
Reference in New Issue
Block a user