初始化项目,由ModelHub XC社区提供模型
Model: RthItalia/NanoLLM-Qwen2.5-14B-v3.1 Source: Original Platform
This commit is contained in:
28
README.md
Normal file
28
README.md
Normal file
@@ -0,0 +1,28 @@
|
||||
---
|
||||
license: other
|
||||
library_name: transformers
|
||||
base_model: Qwen/Qwen2.5-14B-Instruct
|
||||
tags:
|
||||
- nanollm
|
||||
- qwen2.5
|
||||
- safetensors
|
||||
- text-generation
|
||||
---
|
||||
|
||||
# NanoLLM Qwen2.5-14B-Instruct v3.1
|
||||
|
||||
Compact self-contained NanoLLM format is in `nano_compact/`.
|
||||
|
||||
```python
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer
|
||||
repo_id = "RthItalia/NanoLLM-Qwen2.5-14B-v3.1"
|
||||
tokenizer = AutoTokenizer.from_pretrained(repo_id, subfolder="nano_compact", use_fast=True)
|
||||
model = AutoModelForCausalLM.from_pretrained(repo_id, subfolder="nano_compact", trust_remote_code=True, device_map="auto")
|
||||
```
|
||||
|
||||
Validation against 8-bit reference:
|
||||
- avg cosine: 0.98984375
|
||||
- min cosine: 0.9765625
|
||||
- gate: avg >= 0.985
|
||||
|
||||
`nano_compact/model.safetensors` contains Nano quantized tensors and does not require downloading the Qwen base weights.
|
||||
Reference in New Issue
Block a user