Update README.md
This commit is contained in:
22
README.md
22
README.md
@@ -19,3 +19,25 @@ library_name: transformers
|
||||
- **License:** [apache-2.0](https://choosealicense.com/licenses/apache-2.0/)
|
||||
- **Finetuned from model :** [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)
|
||||
- **Dataset:** [sayhan/strix-philosophy-qa](https://huggingface.co/datasets/sayhan/strix-philosophy-qa)
|
||||
---
|
||||
**LoRA rank:** 8
|
||||
**LoRA alpha:** 16
|
||||
**LoRA dropout:** 0
|
||||
**Rank-stabilized LoRA:** Yes
|
||||
**Number of epochs:** 3
|
||||
**Learning rate:** 1e-5
|
||||
**Batch size:** 2
|
||||
**Gradient accumulation steps:** 4
|
||||
**Weight decay:** 0.01
|
||||
**Target modules:**
|
||||
```
|
||||
- Query projection (`q_proj`)
|
||||
- Key projection (`k_proj`)
|
||||
- Value projection (`v_proj`)
|
||||
- Output projection (`o_proj`)
|
||||
- Gate projection (`gate_proj`)
|
||||
- Up projection (`up_proj`)
|
||||
- Down projection (`down_proj`)
|
||||
```
|
||||
|
||||
|
||||
|
||||
Reference in New Issue
Block a user