ModelHub XC 7755b7b7be 初始化项目,由ModelHub XC社区提供模型
Model: prithivMLmods/Carinae-Qwen3-Radiation-4B-GGUF
Source: Original Platform
2026-05-01 19:39:18 +08:00

license, language, base_model, pipeline_tag, library_name, tags
license language base_model pipeline_tag library_name tags
apache-2.0
en
prithivMLmods/Carinae-Qwen3-Radiation-4B
text-generation transformers
text-generation-inference

Carinae-Qwen3-Radiation-4B-GGUF

Carinae-Qwen3-Radiation-4B is a reasoning-focused model fine-tuned on Qwen for Abliterated Reasoning and polished token probabilities, enhancing balanced multilingual generation across mathematics and general-purpose reasoning. It specializes in event-driven logic, structured analysis, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning.

Model Files

File Name Quant Type File Size
Carinae-Qwen3-Radiation-4B.BF16.gguf BF16 8.05 GB
Carinae-Qwen3-Radiation-4B.F16.gguf F16 8.05 GB
Carinae-Qwen3-Radiation-4B.F32.gguf F32 16.1 GB
Carinae-Qwen3-Radiation-4B.Q2_K.gguf Q2_K 1.67 GB
Carinae-Qwen3-Radiation-4B.Q3_K_L.gguf Q3_K_L 2.24 GB
Carinae-Qwen3-Radiation-4B.Q3_K_M.gguf Q3_K_M 2.08 GB
Carinae-Qwen3-Radiation-4B.Q3_K_S.gguf Q3_K_S 1.89 GB
Carinae-Qwen3-Radiation-4B.Q4_K_M.gguf Q4_K_M 2.5 GB
Carinae-Qwen3-Radiation-4B.Q4_K_S.gguf Q4_K_S 2.38 GB
Carinae-Qwen3-Radiation-4B.Q5_K_M.gguf Q5_K_M 2.89 GB
Carinae-Qwen3-Radiation-4B.Q5_K_S.gguf Q5_K_S 2.82 GB
Carinae-Qwen3-Radiation-4B.Q6_K.gguf Q6_K 3.31 GB
Carinae-Qwen3-Radiation-4B.Q8_0.gguf Q8_0 4.28 GB

Quants Usage

(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)

Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):

image.png

Description
Model synced from source: prithivMLmods/Carinae-Qwen3-Radiation-4B-GGUF
Readme 27 KiB