初始化项目,由ModelHub XC社区提供模型
Model: janhq/Jan-v1-4B-GGUF Source: Original Platform
This commit is contained in:
40
.gitattributes
vendored
Normal file
40
.gitattributes
vendored
Normal file
@@ -0,0 +1,40 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
Jan-v1-4B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Jan-v1-4B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Jan-v1-4B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
Jan-v1-4B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
demo.gif filter=lfs diff=lfs merge=lfs -text
|
||||
3
Jan-v1-4B-Q4_K_M.gguf
Normal file
3
Jan-v1-4B-Q4_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:5c083ee3b4b4b099e26299a34b6b9c4a90839aaade6ef35554197ce35dcfd83a
|
||||
size 2497281632
|
||||
3
Jan-v1-4B-Q5_K_M.gguf
Normal file
3
Jan-v1-4B-Q5_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:1a3adf52759386325518978ab7ae1e553978d993311e7a95cce64885cf36c2b9
|
||||
size 2889514592
|
||||
3
Jan-v1-4B-Q6_K.gguf
Normal file
3
Jan-v1-4B-Q6_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:e919cfcff2502a47f62f0ec4c17ac9f5d5916f0bdc5a140be25c1103886c8d45
|
||||
size 3306262112
|
||||
3
Jan-v1-4B-Q8_0.gguf
Normal file
3
Jan-v1-4B-Q8_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:50591c54a3d0e6559083df69a743c622556025a3285e6c5e5177ba4949f70a98
|
||||
size 4280406112
|
||||
91
README.md
Normal file
91
README.md
Normal file
@@ -0,0 +1,91 @@
|
||||
---
|
||||
license: apache-2.0
|
||||
language:
|
||||
- en
|
||||
base_model:
|
||||
- janhq/Jan-v1-4B
|
||||
pipeline_tag: text-generation
|
||||
---
|
||||
# Jan-v1: Advanced Agentic Language Model
|
||||
|
||||
[](https://github.com/menloresearch/deep-research)
|
||||
[](https://opensource.org/licenses/Apache-2.0)
|
||||
[](https://jan.ai/)
|
||||
|
||||
<!-- Optional: If you have a GIF for Jan-v1, include it here like Lucy's. -->
|
||||
<!--  -->
|
||||
|
||||
## Overview
|
||||
|
||||
**Jan-v1** is the first release in the **Jan Family**, designed for agentic reasoning and problem-solving within the [Jan App](https://jan.ai/). Based on our [**Lucy**](https://huggingface.co/Menlo/Lucy) model, Jan-v1 achieves improved performance through model scaling.
|
||||
|
||||
Jan-v1 uses the [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks.
|
||||
|
||||
## Performance
|
||||
|
||||
### Question Answering (SimpleQA)
|
||||
For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy.
|
||||
|
||||

|
||||
|
||||
*The 91.1% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach.*
|
||||
|
||||
### Chat Benchmarks
|
||||
|
||||
These benchmarks evaluate the model's conversational and instructional capabilities.
|
||||
|
||||

|
||||
|
||||
## Quick Start
|
||||
|
||||
### Integration with Jan App
|
||||
|
||||
Jan-v1 is optimized for direct integration with the [Jan App](https://jan.ai/). Simply select the model from the Jan App interface for immediate access to its full capabilities.
|
||||
|
||||

|
||||
|
||||
### Local Deployment
|
||||
|
||||
**Using vLLM:**
|
||||
```bash
|
||||
vllm serve janhq/Jan-v1-4B \
|
||||
--host 0.0.0.0 \
|
||||
--port 1234 \
|
||||
--enable-auto-tool-choice \
|
||||
--tool-call-parser hermes
|
||||
```
|
||||
|
||||
**Using llama.cpp:**
|
||||
```bash
|
||||
llama-server --model jan-v1.gguf \
|
||||
--host 0.0.0.0 \
|
||||
--port 1234 \
|
||||
--jinja \
|
||||
--no-context-shift
|
||||
```
|
||||
|
||||
### Recommended Parameters
|
||||
|
||||
```yaml
|
||||
temperature: 0.6
|
||||
top_p: 0.95
|
||||
top_k: 20
|
||||
min_p: 0.0
|
||||
max_tokens: 2048
|
||||
```
|
||||
|
||||
|
||||
## 🤝 Community & Support
|
||||
|
||||
- **Discussions**: [HuggingFace Community](https://huggingface.co/janhq/Jan-v1-4B/discussions) <!-- Update with your HF model ID -->
|
||||
- **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/)
|
||||
|
||||
## (*) Note
|
||||
By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file [chat_template_raw.jinja](https://huggingface.co/janhq/Jan-v1-4B/blob/main/chat_template_raw.jinja).
|
||||
|
||||
|
||||
## 📄 Citation
|
||||
```bibtex
|
||||
Updated Soon
|
||||
```
|
||||
---
|
||||
Reference in New Issue
Block a user