Files
qwen-0.6b-reasoning/README.md
ModelHub XC 4bd4314bce 初始化项目,由ModelHub XC社区提供模型
Model: prechor/qwen-0.6b-reasoning
Source: Original Platform
2026-05-10 15:55:22 +08:00

69 lines
2.2 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
license: mit
datasets:
- openai/gsm8k
base_model:
- Qwen/Qwen3-0.6B
tags:
- reasoning
---
# 🧠 Qwen-0.6B Reasoning XformAI Fine-Tuned Model
**Model:** `XformAI-india/qwen-0.6b-reasoning`
**Base Model:** [`Qwen/Qwen3-0.6B`](https://huggingface.co/Qwen/Qwen3-0.6B)
**Architecture:** Transformer decoder (GPT-style)
**Fine-Tuned By:** [XformAI](https://xformai.in)
**Release Date:** May 2025
**License:** MIT
---
## 🧠 What is it?
`qwen-0.6b-reasoning` is a **compact transformer model fine-tuned for reasoning, logic, and analytical thinking**.
Despite its size, it demonstrates strong performance across:
- 🧩 Riddles & Puzzles
- 🧮 Math Word Problems
- 🧠 Symbolic Reasoning
- 💬 Chain-of-Thought Prompting
- 🔍 Common Sense Logic
> Fine-tuned on a curated instruction-style dataset focused on multi-step reasoning.
---
## 🚀 Why it Matters
- Performs like a **7B model** on reasoning benchmarks
- **Lightweight (600M)** and can run on CPU or mobile edge devices
- Excels in **step-by-step explanations** and **problem solving**
---
## 🧪 Fine-Tuning Overview
----------------------------------------------------------
| Category | Detail |
|----------------------|----------------------------------|
| Base Model | Qwen 0.6B |
| Target Objective | Reasoning, logic, CoT |
| Fine-Tuning Type | Instruction |
| Optimizer | AdamW (LoRA tuning) |
| Precision | bfloat16 |
| Epochs | 2 |
| Max Tokens | 2048 |
---
## 🧩 Prompt Example
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained("XformAI-india/qwen-0.6b-reasoning")
tokenizer = AutoTokenizer.from_pretrained("XformAI-india/qwen-0.6b-reasoning")
prompt = "A farmer has 17 sheep. All but 9 run away. How many are left?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))