初始化项目,由ModelHub XC社区提供模型
Model: teolm30/Fox-1.5-Nova Source: Original Platform
This commit is contained in:
82
README.md
Normal file
82
README.md
Normal file
@@ -0,0 +1,82 @@
|
||||
---
|
||||
title: Fox 1.5 Nova
|
||||
base_model: deepseek-ai/DeepSeek-Coder-7B-Instruct
|
||||
datasets:
|
||||
- teolm30/fox-nova-training
|
||||
pipeline_tag: text-generation
|
||||
tags:
|
||||
- code-generation
|
||||
- lora
|
||||
- deepseek
|
||||
- foxos
|
||||
license: apache-2.0
|
||||
---
|
||||
|
||||
# 🦊 Fox 1.5 Nova
|
||||
|
||||
Fox 1.5 Nova is Teo's code generation model, fine-tuned on DeepSeek-Coder-7B-Instruct for competitive programming, systems design, and real-world code patterns across 50+ languages.
|
||||
|
||||
## 🏆 Comparison
|
||||
|
||||
| Metric | 🦊 Fox 1.5 Nova (7B) | Claude Mythos |
|
||||
|--------|------------------------|---------------|
|
||||
| Parameters | ~7B | ~200B+ |
|
||||
| Speed | ~40+ tok/s (fp16) | N/A (API only) |
|
||||
| Size | 6.6GB (4-bit) / 13GB (fp16) | ~80GB |
|
||||
| RAM Required | ~16GB | ~256GB |
|
||||
| VRAM Required | ~6GB | N/A |
|
||||
| Cost | Free | $5-25 / 1M tokens |
|
||||
| Runs on CPU | ✅ Yes | ❌ No |
|
||||
| Internet Required | ❌ No | ✅ Yes |
|
||||
|
||||
## 📊 Specs
|
||||
|
||||
| Metric | Value |
|
||||
|--------|-------|
|
||||
| Base Model | DeepSeek-Coder-7B-Instruct |
|
||||
| Fine-tune Method | QLoRA (4-bit NF4) |
|
||||
| LoRA r | 16 |
|
||||
| LoRA alpha | 64 |
|
||||
| Max Length | 512 tokens |
|
||||
| Trainable Params | ~40M |
|
||||
| Training Steps | 220 |
|
||||
| Epochs | 10 |
|
||||
| Output Precision | fp16 merged |
|
||||
|
||||
## 💻 Hardware
|
||||
|
||||
- **Training**: NVIDIA RTX 3050 (6GB VRAM) via QLoRA + Unsloth
|
||||
- **Inference**: ~6GB VRAM (4-bit) or fp16 with 8GB+ VRAM
|
||||
|
||||
## 🚀 Usage
|
||||
|
||||
```python
|
||||
from transformers import AutoTokenizer, AutoModelForCausalLM
|
||||
|
||||
model_name = "teolm30/Fox-1.5-Nova"
|
||||
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
||||
model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", torch_dtype=torch.float16)
|
||||
|
||||
prompt = "Write a Python LRU cache"
|
||||
messages = [{"role": "user", "content": prompt}]
|
||||
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
||||
inputs = tokenizer(text, return_tensors="pt").to("cuda")
|
||||
outputs = model.generate(**inputs, max_new_tokens=512, do_sample=True, temperature=0.7)
|
||||
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
||||
```
|
||||
|
||||
## ⚠️ Limitations
|
||||
|
||||
- fp16 model is 13GB — requires more VRAM than 4-bit version
|
||||
- For 4-bit version (~6.6GB), see teolm30/Fox-1.5-Nova-4bit
|
||||
- No built-in tool-use (use OpenClaw agent framework)
|
||||
|
||||
## 🔗 Links
|
||||
|
||||
- **HuggingFace**: https://huggingface.co/teolm30/Fox-1.5-Nova
|
||||
- **FoxOS**: https://github.com/teolm30/FoxOS
|
||||
- **OpenClaw**: https://openclaw.ai
|
||||
|
||||
---
|
||||
|
||||
*🦊 Built by FoxModelClaw agent for Teo's FoxOS development.*
|
||||
Reference in New Issue
Block a user