Update README.md
This commit is contained in:
@@ -10,7 +10,9 @@ tags:
|
|||||||
- llama
|
- llama
|
||||||
---
|
---
|
||||||
<img src="theprint_18b_moe.png" width="420" />
|
<img src="theprint_18b_moe.png" width="420" />
|
||||||
|
|
||||||
# theprint-MoE-8x3-0126-GGUF
|
# theprint-MoE-8x3-0126-GGUF
|
||||||
|
|
||||||
An 18B parameter Mixture of Experts model combining 8 specialized 3B experts, with 2 experts activated per token by default (configurable up to 4 at inference).
|
An 18B parameter Mixture of Experts model combining 8 specialized 3B experts, with 2 experts activated per token by default (configurable up to 4 at inference).
|
||||||
|
|
||||||
## Architecture
|
## Architecture
|
||||||
|
|||||||
Reference in New Issue
Block a user