初始化项目,由ModelHub XC社区提供模型

Model: OrionLLM/Nebula
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-11 20:17:03 +08:00
commit 21bfa7e3bb
9 changed files with 327706 additions and 0 deletions

39
.gitattributes vendored Normal file
View File

@@ -0,0 +1,39 @@
*.7z filter=lfs diff=lfs merge=lfs -text
*.arrow filter=lfs diff=lfs merge=lfs -text
*.bin filter=lfs diff=lfs merge=lfs -text
*.bz2 filter=lfs diff=lfs merge=lfs -text
*.ckpt filter=lfs diff=lfs merge=lfs -text
*.ftz filter=lfs diff=lfs merge=lfs -text
*.gz filter=lfs diff=lfs merge=lfs -text
*.h5 filter=lfs diff=lfs merge=lfs -text
*.joblib filter=lfs diff=lfs merge=lfs -text
*.lfs.* filter=lfs diff=lfs merge=lfs -text
*.mlmodel filter=lfs diff=lfs merge=lfs -text
*.model filter=lfs diff=lfs merge=lfs -text
*.msgpack filter=lfs diff=lfs merge=lfs -text
*.npy filter=lfs diff=lfs merge=lfs -text
*.npz filter=lfs diff=lfs merge=lfs -text
*.onnx filter=lfs diff=lfs merge=lfs -text
*.ot filter=lfs diff=lfs merge=lfs -text
*.parquet filter=lfs diff=lfs merge=lfs -text
*.pb filter=lfs diff=lfs merge=lfs -text
*.pickle filter=lfs diff=lfs merge=lfs -text
*.pkl filter=lfs diff=lfs merge=lfs -text
*.pt filter=lfs diff=lfs merge=lfs -text
*.pth filter=lfs diff=lfs merge=lfs -text
*.rar filter=lfs diff=lfs merge=lfs -text
*.safetensors filter=lfs diff=lfs merge=lfs -text
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.tar.* filter=lfs diff=lfs merge=lfs -text
*.tar filter=lfs diff=lfs merge=lfs -text
*.tflite filter=lfs diff=lfs merge=lfs -text
*.tgz filter=lfs diff=lfs merge=lfs -text
*.wasm filter=lfs diff=lfs merge=lfs -text
*.xz filter=lfs diff=lfs merge=lfs -text
*.zip filter=lfs diff=lfs merge=lfs -text
*.zst filter=lfs diff=lfs merge=lfs -text
*tfevents* filter=lfs diff=lfs merge=lfs -text
figures/baguettotron_structure.png filter=lfs diff=lfs merge=lfs -text
figures/table_evaluation.png filter=lfs diff=lfs merge=lfs -text
figures/training_baguettotron.png filter=lfs diff=lfs merge=lfs -text
figures/comparison_models.png filter=lfs diff=lfs merge=lfs -text

71
README.md Normal file
View File

@@ -0,0 +1,71 @@
---
license: apache-2.0
pipeline_tag: text-generation
library_name: transformers
tags:
- nebula
- reasoning
- text-generation
- transformers
---
# Nebula
<p align="center">
<img src="https://cdn-uploads.huggingface.co/production/uploads/685ea8ff7b4139b6845ce395/YF0kEDYMGJhcM3Lbl2EOD.png" alt="Nebula logo" width="100">
</p>
## 1. Introduction
**Nebula** is a **320M-parameter** generalist Small Reasoning Model trained on **200B+ tokens**, designed for edge AI and on-device deployment.
Nebula is designed to deliver an unusually strong balance of **memory**, **general reasoning**, **math**, and **retrieval-friendly behavior** for its size class, aiming to outperform many small models of a similar parameter range on non-code, industry-style benchmarks.
## 2. Reasoning style
Nebulas reasoning traces use an intentionally compact style with **dense, short, frequently non-verbal sentences**, optimized for efficiency under limited model capacity.
Traces use the following stenographic notation integrated into special tokens:
### Logical markers
| Token | Meaning | Usage |
| ----- | ------- | ----- |
| **→** | derivation / implication | For very short causal/logical flow |
| **↺** | iterative return / refinement loop | For backtracking, reconsidering priors, RAG re-querying |
| **?** | uncertainty/questions to resolve | Can be appended to short expressions/words, not only interrogatives |
| **!/※** | insight/breakthroughs | Emphatic mark for knowledge discovery |
| **≈** | approximation/estimates | For intermediary hypothesis / uncertain preliminary statements |
| **∴** | therefore / final step | Use sparingly to mark stable conclusions |
### Uncertainty
| Token | Meaning | Usage |
| ----- | ------- | ----- |
| **●** | high confidence | well-supported empirical/theoretical ground; “anchor points.” |
| **◐** | medium/partial confidence | incomplete data; plausible but unverified links |
| **○** | low confidence | speculation, missing context, weak inference chain |
| **⚠** | bias/premise risk | domain mismatch, cultural assumptions, language-switch artifacts |
| **?maybe?** | soft speculation | marks tentative ideas, branches that might collapse later |
### Verification process
| Token | Meaning | Usage |
| ----- | ------- | ----- |
| **☐** | unverified hypothesis | raw claim, no cross-check yet |
| **☑** | intermediate verification | one source/argument supports it |
| **✓** | confirmed/validated | multiple independent supports (●-level) |
This reasoning format is designed to remain expressive while being lightweight enough for a small model.
## 3. Fine-Tuning/RL
Nebula has been successfully fine-tuned for a variety of tasks
Because Nebula is a reasoning-oriented model, it is expected to train well with reinforcement learning methods such as **GRPO**, both for **verifiable tasks** (with objective rewards) and for subjective tasks using an **LLM-as-a-judge**.
## 4. Benchmarks
| Model | MMLU |
|------|-----:|
| **Nebula** | **40.0** |
| SmolLM2-360M | 35.8 |
| Gemma 3 270M (IT) | 26.5 |
| Granite-4.0-H-350M | 36.21 |

7
chat_template.json Normal file
View File

@@ -0,0 +1,7 @@
{
"chat_template": "{% for m in messages %}<|im_start|>{{ m['role'] }}\n{{ m['content'] }}<|im_end|>\n{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n<think>\n{% endif %}",
"eos_token": "<|im_end|>",
"bos_token": "<|im_start|>",
"stop": ["<|im_end|>"],
"roles": { "user": "user", "assistant": "assistant", "system": "system" }
}

29
config.json Normal file
View File

@@ -0,0 +1,29 @@
{
"architectures": [
"LlamaForCausalLM"
],
"attention_bias": false,
"attention_dropout": 0.0,
"bos_token_id": 1,
"eos_token_id": 2,
"head_dim": 64,
"hidden_act": "silu",
"hidden_size": 576,
"initializer_range": 0.02,
"intermediate_size": 1536,
"max_position_embeddings": 4096,
"mlp_bias": false,
"model_type": "llama",
"num_attention_heads": 9,
"num_hidden_layers": 80,
"num_key_value_heads": 3,
"pretraining_tp": 1,
"rms_norm_eps": 1e-05,
"rope_scaling": null,
"rope_theta": 10000,
"tie_word_embeddings": true,
"torch_dtype": "bfloat16",
"transformers_version": "4.51.3",
"use_cache": true,
"vocab_size": 65536
}

6
generation_config.json Normal file
View File

@@ -0,0 +1,6 @@
{
"_from_model_config": true,
"bos_token_id": 1,
"eos_token_id": 2,
"transformers_version": "4.51.3"
}

3
model.safetensors Normal file
View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:dbb3fe0fd0d97a28c140aa315ec4a651f20432e9b7a509908a620190f506644b
size 641995416

53
special_tokens_map.json Normal file
View File

@@ -0,0 +1,53 @@
{
"additional_special_tokens": [
"<|im_start|>",
"<|im_end>",
"<think>",
"</think>",
"source_1",
"source_2",
"source_3",
"source_4",
"source_5",
"source_6",
"source_7",
"source_8",
"source_9",
"source_10",
"<ref",
"</ref>",
"→",
"↺",
"※",
"?maybe?",
"●",
"◐",
"○",
"⚠",
"☐",
"☑",
"✓",
"⟨H≈0.1⟩",
"⟨H≈0.2⟩",
"⟨H≈0.3⟩",
"⟨H≈0.4⟩",
"⟨H≈0.5⟩",
"⟨H≈0.6⟩",
"⟨H≈0.7⟩",
"⟨H≈0.8⟩",
"⟨H≈0.9⟩",
"⟨H≈1.0⟩",
"⟨H≈1.1⟩",
"⟨H≈1.2⟩",
"⟨H≈1.3⟩",
"⟨H≈1.4⟩",
"⟨H≈1.5⟩",
"⟨H≈1.6⟩",
"⟨H≈1.7⟩",
"⟨H≈1.8⟩"
],
"bos_token": "<|begin_of_text|>",
"eos_token": "<|end_of_text|>",
"pad_token": "[PAD]",
"unk_token": "[UNK]"
}

327047
tokenizer.json Normal file

File diff suppressed because it is too large Load Diff

451
tokenizer_config.json Normal file
View File

@@ -0,0 +1,451 @@
{
"added_tokens_decoder": {
"0": {
"content": "[UNK]",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"1": {
"content": "<|begin_of_text|>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"2": {
"content": "<|end_of_text|>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"3": {
"content": "[PAD]",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65491": {
"content": "<|im_start|>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65492": {
"content": "<|im_end>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65493": {
"content": "<think>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65494": {
"content": "</think>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65495": {
"content": "source_1",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65496": {
"content": "source_2",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65497": {
"content": "source_3",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65498": {
"content": "source_4",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65499": {
"content": "source_5",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65500": {
"content": "source_6",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65501": {
"content": "source_7",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65502": {
"content": "source_8",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65503": {
"content": "source_9",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65504": {
"content": "source_10",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65505": {
"content": "<ref",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65506": {
"content": "</ref>",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65507": {
"content": "→",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65508": {
"content": "↺",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65509": {
"content": "※",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65510": {
"content": "?maybe?",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65511": {
"content": "●",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65512": {
"content": "◐",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65513": {
"content": "○",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65514": {
"content": "⚠",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65515": {
"content": "☐",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65516": {
"content": "☑",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65517": {
"content": "✓",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65518": {
"content": "⟨H≈0.1⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65519": {
"content": "⟨H≈0.2⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65520": {
"content": "⟨H≈0.3⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65521": {
"content": "⟨H≈0.4⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65522": {
"content": "⟨H≈0.5⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65523": {
"content": "⟨H≈0.6⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65524": {
"content": "⟨H≈0.7⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65525": {
"content": "⟨H≈0.8⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65526": {
"content": "⟨H≈0.9⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65527": {
"content": "⟨H≈1.0⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65528": {
"content": "⟨H≈1.1⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65529": {
"content": "⟨H≈1.2⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65530": {
"content": "⟨H≈1.3⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65531": {
"content": "⟨H≈1.4⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65532": {
"content": "⟨H≈1.5⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65533": {
"content": "⟨H≈1.6⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65534": {
"content": "⟨H≈1.7⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
},
"65535": {
"content": "⟨H≈1.8⟩",
"lstrip": false,
"normalized": false,
"rstrip": false,
"single_word": false,
"special": true
}
},
"additional_special_tokens": [
"<|im_start|>",
"<|im_end>",
"<think>",
"</think>",
"source_1",
"source_2",
"source_3",
"source_4",
"source_5",
"source_6",
"source_7",
"source_8",
"source_9",
"source_10",
"<ref",
"</ref>",
"→",
"↺",
"※",
"?maybe?",
"●",
"◐",
"○",
"⚠",
"☐",
"☑",
"✓",
"⟨H≈0.1⟩",
"⟨H≈0.2⟩",
"⟨H≈0.3⟩",
"⟨H≈0.4⟩",
"⟨H≈0.5⟩",
"⟨H≈0.6⟩",
"⟨H≈0.7⟩",
"⟨H≈0.8⟩",
"⟨H≈0.9⟩",
"⟨H≈1.0⟩",
"⟨H≈1.1⟩",
"⟨H≈1.2⟩",
"⟨H≈1.3⟩",
"⟨H≈1.4⟩",
"⟨H≈1.5⟩",
"⟨H≈1.6⟩",
"⟨H≈1.7⟩",
"⟨H≈1.8⟩"
],
"bos_token": "<|begin_of_text|>",
"clean_up_tokenization_spaces": true,
"eos_token": "<|end_of_text|>",
"extra_special_tokens": {},
"model_max_length": 1000000000000000019884624838656,
"pad_token": "[PAD]",
"tokenizer_class": "PreTrainedTokenizer",
"unk_token": "[UNK]"
}