初始化项目,由ModelHub XC社区提供模型

Model: khazarai/Quantum-ToT
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-05-02 13:04:27 +08:00
commit 8698c423ea
12 changed files with 152016 additions and 0 deletions

110
README.md Normal file
View File

@@ -0,0 +1,110 @@
---
library_name: transformers
tags:
- Physics
- Quantum
- reasoning
- unsloth
- sft
- lora
license: apache-2.0
datasets:
- mattwesney/CoT_Reasoning_Quantom_Physics_And_Computing
language:
- en
base_model:
- unsloth/Qwen3-1.7B
pipeline_tag: text-generation
---
# Quantum-ToT
## Model Details
Quantum-ToT is a fine-tuned variant of Qwen3-1.7B, optimized for Chain-of-Thought (CoT) reasoning in quantum mechanics and quantum computing contexts.
This model was trained using the [moremilk/CoT_Reasoning_Quantum_Physics_And_Computing](https://huggingface.co/datasets/moremilk/CoT_Reasoning_Quantom_Physics_And_Computing) dataset — a curated collection of questionanswer pairs that go beyond surface-level definitions to show the logical reasoning process behind quantum concepts.
The goal of this fine-tuning is to enhance the models ability to:
- Explain quantum principles with structured, step-by-step logic
- Reason through conceptual problems in quantum physics and computing
- Support educational and research applications that require interpretable reasoning chains
## Uses
### Direct Use
- Educational assistance in quantum physics and quantum computing
- AI tutors or reasoning assistants for STEM learning
- Conceptual reasoning benchmarks involving quantum phenomena
- Research in reasoning-aware model behavior and CoT interpretability
### Out of Scope
- Predicting new or unverified physical phenomena
- Running quantum simulations or algorithmic derivations
- Hardware-level quantum design
- Real-time physics predictions
## Bias, Risks, and Limitations
- May hallucinate if prompted outside the quantum domain
- Not suitable for advanced quantum algorithm design or experimental predictions
## How to Get Started with the Model
Use the code below to get started with the model.
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("khazarai/Quantum-ToT")
model = AutoModelForCausalLM.from_pretrained(
"khazarai/Quantum-ToT",
device_map={"": 0}
)
question = """
Explain the Heisenberg Uncertainty Principle in detail, including its mathematical formulation, physical implications, and common misconceptions.
"""
messages = [
{"role" : "user", "content" : question}
]
text = tokenizer.apply_chat_template(
messages,
tokenize = False,
add_generation_prompt = True,
enable_thinking = True,
)
from transformers import TextStreamer
_ = model.generate(
**tokenizer(text, return_tensors = "pt").to("cuda"),
max_new_tokens = 3000,
temperature = 0.6,
top_p = 0.95,
top_k = 20,
streamer = TextStreamer(tokenizer, skip_prompt = True),
)
```
### Dataset
Dataset: [moremilk/CoT_Reasoning_Quantum_Physics_And_Computing](https://huggingface.co/datasets/moremilk/CoT_Reasoning_Quantom_Physics_And_Computing)
This dataset contains rich reasoning-based questionanswer pairs covering:
- Core quantum principles: superposition, entanglement, measurement
- Effects of quantum gates (Hadamard, Pauli-X/Y/Z, etc.) on qubits
- Multi-qubit reasoning (e.g., Bell states, entangled systems)
- Basic quantum algorithms and logical operations
- Probabilistic interpretation of measurement outcomes
Each entry includes:
- think block → models internal reasoning process
- answer block → final concise explanation or solution
The dataset focuses on conceptual understanding rather than heavy mathematical derivations or complex quantum hardware design.