Sayhan Yalvaçer d82f3d8e4b Update README.md
2024-02-18 15:53:23 +00:00
2024-02-18 08:36:46 +00:00
2024-02-18 08:36:46 +00:00
2024-02-18 08:40:15 +00:00
2024-02-18 12:48:58 +00:00
2024-02-18 15:53:23 +00:00
2024-02-18 08:40:15 +00:00
2024-02-18 08:40:15 +00:00
2024-02-18 08:40:15 +00:00
2024-02-18 08:40:15 +00:00

language, license, tags, base_model, datasets, library_name
language license tags base_model datasets library_name
en
apache-2.0
trl
text-generation-inference
unsloth
mistral
gguf
teknium/OpenHermes-2.5-Mistral-7B
sayhan/strix-philosophy-qa
transformers

image/png

OpenHermes 2.5 Stix Philosophy Mistral 7B


LoRA rank: 8
LoRA alpha: 16
LoRA dropout: 0
Rank-stabilized LoRA: Yes
Number of epochs: 3
Learning rate: 1e-5
Batch size: 2
Gradient accumulation steps: 4
Weight decay: 0.01
Target modules:

  - Query projection (`q_proj`)
  - Key projection (`k_proj`)
  - Value projection (`v_proj`)
  - Output projection (`o_proj`)
  - Gate projection (`gate_proj`)
  - Up projection (`up_proj`)
  - Down projection (`down_proj`)
Description
Model synced from source: sayhan/OpenHermes-2.5-Strix-Philosophy-Mistral-7B-LoRA
Readme 541 KiB