初始化项目,由ModelHub XC社区提供模型
Model: CorticalStack/pastiche-crown-clown-7b-dare-dpo Source: Original Platform
This commit is contained in:
29
.ipynb_checkpoints/README-checkpoint.md
Normal file
29
.ipynb_checkpoints/README-checkpoint.md
Normal file
@@ -0,0 +1,29 @@
|
||||
---
|
||||
license: apache-2.0
|
||||
tags:
|
||||
- dpo
|
||||
base_model:
|
||||
- CorticalStack/pastiche-crown-clown-7b-dare
|
||||
dataset:
|
||||
- jondurbin/truthy-dpo-v0.1
|
||||
---
|
||||
|
||||
# CorticalStack/pastiche-crown-clown-7b-dare-dpo
|
||||
|
||||
CorticalStack/pastiche-crown-clown-7b-dare-dpo is a DPO fine-tuned version of [CorticalStack/pastiche-crown-clown-7b-dare](https://huggingface.co/CorticalStack/pastiche-crown-clown-7b-dare) using the [jondurbin/truthy-dpo-v0.1](https://huggingface.co/datasets/jondurbin/truthy-dpo-v0.1) dataset.
|
||||
|
||||
### LoRA
|
||||
- r: 16
|
||||
- LoRA alpha: 16
|
||||
- LoRA dropout: 0.05
|
||||
|
||||
### Training arguments
|
||||
- Batch size: 4
|
||||
- Gradient accumulation steps: 4
|
||||
- Optimizer: paged_adamw_32bit
|
||||
- Max steps: 200
|
||||
- Learning rate: 5e-05
|
||||
- Learning rate scheduler type: cosine
|
||||
- Beta: 0.1
|
||||
- Max prompt length: 1024
|
||||
- Max length: 1536
|
||||
Reference in New Issue
Block a user