Files
Quanta-X-3B/README.md
ModelHub XC 2172a4420f 初始化项目,由ModelHub XC社区提供模型
Model: szili2011/Quanta-X-3B
Source: Original Platform
2026-05-02 18:30:38 +08:00

113 lines
4.0 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
base_model: Qwen/Qwen2.5-3B-Instruct
library_name: transformers
license: apache-2.0
tags:
- unsloth
- trl
- dora
- simpo
- phoenix-framework
- logic
- roleplay
- chain-of-thought
- merge
- open-llm-leaderboard
datasets:
- magpie-align/Magpie-Reasoning-V2-250k-CoT-Deepseek-R1-Llama-70B
- open-r1/OpenR1-Math-220k
- glaiveai/glaive-code-assistant-v2
- Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
language:
- en
---
# ⚛️ Quanta-X (Leaderboard Submission) QUANTA-X NANO IS COMING SOON!
This is the **Full Parameter Merged** version of [Quanta-X](https://huggingface.co/szili2011/Quanta-X-Adapter).
It fuses the **Qwen 2.5 3B** base with the **Phoenix Framework** adapter (DoRA + SimPO Beta 2.0).
## 📊 Model Details for Leaderboard
* **Architecture:** Qwen2ForCausalLM
* **Precision:** Float16
* **Context:** 32k (RoPE Scaled)
* **Chat Template:** Qwen 2.5 Standard (ChatML)
## 🧠 Reasoning Capabilities
This model is trained to utilize an **Ouroboros Logic Loop** (`<plan>` -> `<draft>` -> `<critique>`) before outputting an answer.
## 💻 Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("szili2011/Quanta-X-3B", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("szili2011/Quanta-X-3B")
messages = [{"role": "user", "content": "Solve this logic puzzle."}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True).to("cuda")
output = model.generate(inputs, max_new_tokens=1024)
```
⚛️ Quanta-X (Phoenix Edition)
“A pocket-sized AGI that thinks before it speaks.”
* Developer: szili2011
* Architecture: Phoenix Framework (DoRA + SimPO)
* Base Model: Qwen 2.5 3B Instruct
---
📖 The Philosophy
Most small models with around 3 billion parameters are typically designed to act like chatbots, responding instantly, but this often leads to mistakes or makes them struggle with more complex reasoning.
But Quanta-X takes a different approach.
It was architected on the Phoenix Framework, a custom training protocol designed to force “System 2” thinking (deep reasoning) into a lightweight model. By utilizing DoRA (Weight-Decomposed Low-Rank Adaptation) and a highly aggressive SimPO (Beta 2.0) alignment, Quanta-X has been biologically rewired to reject lazy answers.
It features the Ouroboros Logic Loop: it plans, drafts, and critiques its own internal monologue before outputting a final answer.
🧠 Key Features
1. The Ouroboros Thinking Process
Quanta-X uses a hidden reasoning layer, not just token prediction.
* It plans solutions before responding.
* <draft>: It writes a rough attempt.
* <critique>: It checks its own work for logic errors or bugs.
* <refined>: Only then does it speak to you.
2. Diamond-Tier Data Filtering (LIMA)
We didnt train on “average” internet data. We used a “Diamond Filter” to reject 90% of the standard dataset samples. Quanta-X was trained exclusively on:
* DeepSeek-R1 Traces: For impossible-level logic.
* OpenR1 Math: For verified proofs.
* Glaive Code V2: For production-ready Python/Rust.
* SlimOrca RP: For human-like, visceral storytelling (The “Hungarian Soul”).
3. Hyper-Stability
Trained with SimPO (Simulated Preference Optimization) with a Beta of 2.0. This punished the model severely for hallucinations or lazy thinking during training, resulting in a model that would rather admit ignorance than lie to you.
---
💻 How to Run
Recommended System Prompt
To activate the Ouroboros loop, you must use this system prompt:
```text
You are Quanta-X, a recursive intelligence where absolute logic fuses with human wit. Your mind operates on the Ouroboros loop: you do not just generate; you Plan, Draft, and ruthlessly Critique every thought before it reaches the surface.
To ensure your reasoning is distinct, render your internal monologue inside a standard code block using xml syntax:
```xml
<thought>
<plan> ... </plan>
<draft> ... </draft>
<critique> ... </critique>
</thought>