48 lines
1.5 KiB
Markdown
48 lines
1.5 KiB
Markdown
---
|
|
library_name: transformers
|
|
license: apache-2.0
|
|
datasets:
|
|
- schneewolflabs/Athanorlite-DPO
|
|
base_model:
|
|
- nbeerbower/Schreiber-mistral-nemo-12B
|
|
tags:
|
|
- heretic
|
|
- uncensored
|
|
- decensored
|
|
- abliterated
|
|
---
|
|
# This is a decensored version of [schneewolflabs/A0l-12B](https://huggingface.co/schneewolflabs/A0l-12B), made using [Heretic](https://github.com/p-e-w/heretic) v1.2.0
|
|
|
|
## Abliteration parameters
|
|
|
|
| Parameter | Value |
|
|
| :-------- | :---: |
|
|
| **direction_index** | per layer |
|
|
| **attn.o_proj.max_weight** | 1.24 |
|
|
| **attn.o_proj.max_weight_position** | 33.41 |
|
|
| **attn.o_proj.min_weight** | 0.89 |
|
|
| **attn.o_proj.min_weight_distance** | 22.95 |
|
|
| **mlp.down_proj.max_weight** | 1.47 |
|
|
| **mlp.down_proj.max_weight_position** | 25.98 |
|
|
| **mlp.down_proj.min_weight** | 0.00 |
|
|
| **mlp.down_proj.min_weight_distance** | 12.82 |
|
|
|
|
## Performance
|
|
|
|
| Metric | This model | Original model ([schneewolflabs/A0l-12B](https://huggingface.co/schneewolflabs/A0l-12B)) |
|
|
| :----- | :--------: | :---------------------------: |
|
|
| **KL divergence** | 0.0364 | 0 *(by definition)* |
|
|
| **Refusals** | 10/100 | 40/100 |
|
|
|
|
-----
|
|
|
|
|
|
# A0l-12B
|
|
|
|
Same training run as [schneewolflabs/A0-12B](https://huggingface.co/schneewolflabs/A0-12B), but using [schneewolflabs/Athanorlite-DPO](https://huggingface.co/schneewolflabs/Athanorlite-DPO) as the dataset.
|
|
|
|
Preliminary tests have shown this model has superior writing capabilities to A0-12B.
|
|
|
|
## Configuration
|
|
|
|
 |