commit 47a84523a41b305b0a18a9957435c8149e83b580 Author: ModelHub XC Date: Fri Apr 10 12:52:00 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: janhq/Jan-v1-4B-GGUF Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..a7cf9ac --- /dev/null +++ b/.gitattributes @@ -0,0 +1,40 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +Jan-v1-4B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Jan-v1-4B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +Jan-v1-4B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +Jan-v1-4B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +demo.gif filter=lfs diff=lfs merge=lfs -text diff --git a/Jan-v1-4B-Q4_K_M.gguf b/Jan-v1-4B-Q4_K_M.gguf new file mode 100644 index 0000000..47e0a4b --- /dev/null +++ b/Jan-v1-4B-Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c083ee3b4b4b099e26299a34b6b9c4a90839aaade6ef35554197ce35dcfd83a +size 2497281632 diff --git a/Jan-v1-4B-Q5_K_M.gguf b/Jan-v1-4B-Q5_K_M.gguf new file mode 100644 index 0000000..98b357e --- /dev/null +++ b/Jan-v1-4B-Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a3adf52759386325518978ab7ae1e553978d993311e7a95cce64885cf36c2b9 +size 2889514592 diff --git a/Jan-v1-4B-Q6_K.gguf b/Jan-v1-4B-Q6_K.gguf new file mode 100644 index 0000000..0a135b9 --- /dev/null +++ b/Jan-v1-4B-Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e919cfcff2502a47f62f0ec4c17ac9f5d5916f0bdc5a140be25c1103886c8d45 +size 3306262112 diff --git a/Jan-v1-4B-Q8_0.gguf b/Jan-v1-4B-Q8_0.gguf new file mode 100644 index 0000000..9d14df2 --- /dev/null +++ b/Jan-v1-4B-Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50591c54a3d0e6559083df69a743c622556025a3285e6c5e5177ba4949f70a98 +size 4280406112 diff --git a/README.md b/README.md new file mode 100644 index 0000000..9eef74b --- /dev/null +++ b/README.md @@ -0,0 +1,91 @@ +--- +license: apache-2.0 +language: +- en +base_model: +- janhq/Jan-v1-4B +pipeline_tag: text-generation +--- +# Jan-v1: Advanced Agentic Language Model + +[![GitHub](https://img.shields.io/badge/GitHub-Repository-blue?logo=github)](https://github.com/menloresearch/deep-research) +[![License](https://img.shields.io/badge/License-Apache%202.0-yellow)](https://opensource.org/licenses/Apache-2.0) +[![Jan App](https://img.shields.io/badge/Powered%20by-Jan%20App-purple?style=flat&logo=android)](https://jan.ai/) + + + + +## Overview + +**Jan-v1** is the first release in the **Jan Family**, designed for agentic reasoning and problem-solving within the [Jan App](https://jan.ai/). Based on our [**Lucy**](https://huggingface.co/Menlo/Lucy) model, Jan-v1 achieves improved performance through model scaling. + +Jan-v1 uses the [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks. + +## Performance + +### Question Answering (SimpleQA) +For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy. + +![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/1_9-r0GpBACSsWoUVZ1Xg.png) + +*The 91.1% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach.* + +### Chat Benchmarks + +These benchmarks evaluate the model's conversational and instructional capabilities. + +![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/f3bzNYRuA_iTFQIcvu6Rr.png) + +## Quick Start + +### Integration with Jan App + +Jan-v1 is optimized for direct integration with the [Jan App](https://jan.ai/). Simply select the model from the Jan App interface for immediate access to its full capabilities. + +![image/gif](demo.gif) + +### Local Deployment + +**Using vLLM:** +```bash +vllm serve janhq/Jan-v1-4B \ + --host 0.0.0.0 \ + --port 1234 \ + --enable-auto-tool-choice \ + --tool-call-parser hermes +``` + +**Using llama.cpp:** +```bash +llama-server --model jan-v1.gguf \ + --host 0.0.0.0 \ + --port 1234 \ + --jinja \ + --no-context-shift +``` + +### Recommended Parameters + +```yaml +temperature: 0.6 +top_p: 0.95 +top_k: 20 +min_p: 0.0 +max_tokens: 2048 +``` + + +## 🤝 Community & Support + +- **Discussions**: [HuggingFace Community](https://huggingface.co/janhq/Jan-v1-4B/discussions) +- **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/) + +## (*) Note +By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file [chat_template_raw.jinja](https://huggingface.co/janhq/Jan-v1-4B/blob/main/chat_template_raw.jinja). + + +## 📄 Citation +```bibtex +Updated Soon +``` +--- \ No newline at end of file diff --git a/demo.gif b/demo.gif new file mode 100644 index 0000000..2f12f71 --- /dev/null +++ b/demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd99f33772b20b384a74efc5c386863a1f969651828f477b73b446583df5b0af +size 170956325