Update README.md
This commit is contained in:
@@ -10,7 +10,9 @@ tags:
|
||||
- llama
|
||||
---
|
||||
<img src="theprint_18b_moe.png" width="420" />
|
||||
|
||||
# theprint-MoE-8x3-0126-GGUF
|
||||
|
||||
An 18B parameter Mixture of Experts model combining 8 specialized 3B experts, with 2 experts activated per token by default (configurable up to 4 at inference).
|
||||
|
||||
## Architecture
|
||||
|
||||
Reference in New Issue
Block a user