base_model, tags, language, license, datasets
base_model tags language license datasets
N-Bot-Int/MistThena7B-V2
text-generation-inference
transformers
mistral
rp
gguf
en
apache-2.0
N-Bot-Int/Iris-Uncensored-R2
N-Bot-Int/Millie-R1_DPO
N-Bot-Int/Millia-R1_DPO

Support Us Through

image/png

GGUF Version

GGUF with Quants! Allowing you to run models using KoboldCPP and other AI Environments!

Quantizations:

Quant Type Benefits Cons
Q4_K_M Smallest size (fastest inference) Lowest accuracy compared to other quants
Requires the least VRAM/RAM May struggle with complex reasoning
Ideal for edge devices & low-resource setups Can produce slightly degraded text quality
Q5_K_M Better accuracy than Q4, while still compact Slightly larger model size than Q4
Good balance between speed and precision Needs a bit more VRAM than Q4
Works well on mid-range GPUs Still not as accurate as higher-bit models
Q8_0 Highest accuracy (closest to full model) Requires significantly more VRAM/RAM
Best for complex reasoning & detailed outputs Slower inference compared to Q4 & Q5
Suitable for high-end GPUs & serious workloads Larger file size (takes more storage)

Model Details:

Read the Model details on huggingface Model Detail Here!

Description
Model synced from source: N-Bot-Int/MistThena7BV2-GGUF
Readme 25 KiB