Files
ModelHub XC 0da03965b1 初始化项目,由ModelHub XC社区提供模型
Model: DJLougen/MiroThinker-1.7-mini-GGUF-Q8_0
Source: Original Platform
2026-04-11 04:10:56 +08:00

1.4 KiB

license, base_model, tags, language
license base_model tags language
apache-2.0 miromind-ai/MiroThinker-1.7-mini
gguf
quantized
qwen3_moe
text-generation
agent
deep-research
en

MiroThinker-1.7-mini GGUF Q8_0

Q8_0 GGUF quantization of miromind-ai/MiroThinker-1.7-mini.

Model Details

  • Original Model: miromind-ai/MiroThinker-1.7-mini (Qwen3 MoE, 30.5B params)
  • Quantization: Q8_0 (8-bit)
  • File Size: ~31 GB
  • Format: GGUF (llama.cpp compatible)
  • Max Context: 256K tokens
  • Max Tool Calls: 300

About MiroThinker-1.7-mini

MiroThinker-1.7-mini is a deep research agent model fine-tuned from Qwen3-30B-A3B-Thinking-2507. It achieves state-of-the-art performance in deep research tasks among open-source models.

Benchmarks (original BF16)

Benchmark Score
BrowseComp 74.0%
BrowseComp-ZH 75.3% (SOTA)
GAIA-Val-165 82.7%
HLE-Text 42.9%

Usage

Works with any GGUF-compatible runtime: llama.cpp, Ollama, LM Studio, etc.

Ollama:

ollama run hf.co/DJLougen/MiroThinker-1.7-mini-GGUF-Q8_0

llama.cpp:

llama-cli -m MiroThinker-1.7-mini-Q8_0.gguf -c 8192 -n 512
  • temperature: 1.0
  • top_p: 0.95
  • repetition_penalty: 1.05

Credits