Compare commits

..

10 Commits

Author SHA1 Message Date
nguyen hoang thuan
a567ac3bab Update README.md 2025-08-13 15:38:40 +00:00
Jan (Homebrew Research)
84155fa599 Update README.md 2025-08-12 11:34:09 +00:00
Jan (Homebrew Research)
1c3450457b Update README.md 2025-08-12 08:45:18 +00:00
nguyen hoang thuan
c3d2619969 Update README.md 2025-08-12 04:30:37 +00:00
Jan (Homebrew Research)
cae60a05d1 Update README.md 2025-08-12 03:22:35 +00:00
Jan (Homebrew Research)
654351c1e4 Update README.md 2025-08-12 03:01:31 +00:00
Jan (Homebrew Research)
1965229259 Update README.md 2025-08-12 02:15:42 +00:00
Jan (Homebrew Research)
55f2815b87 Update README.md 2025-08-12 02:11:15 +00:00
Jan (Homebrew Research)
18e9c48c70 Upload demo.gif 2025-08-12 01:56:47 +00:00
Jan (Homebrew Research)
57934a8b26 Delete demo.gif 2025-08-12 01:56:15 +00:00
2 changed files with 18 additions and 12 deletions

View File

@@ -3,7 +3,7 @@ license: apache-2.0
language:
- en
base_model:
- Qwen/Qwen3-4B-Thinking-2507
- janhq/Jan-v1-4B
pipeline_tag: text-generation
---
# Jan-v1: Advanced Agentic Language Model
@@ -17,24 +17,24 @@ pipeline_tag: text-generation
## Overview
Introducing **Jan-v1**, the first release in the **Jan Family** specifically designed for advanced agentic reasoning and complex problem-solving within the [Jan App](https://jan.ai/). Building on the innovative agentic capabilities of our earlier [**Lucy** ](https://huggingface.co/Menlo/Lucy) model, Jan-v1 represents a significant leap forward through strategic model scaling.
**Jan-v1** is the first release in the **Jan Family**, designed for agentic reasoning and problem-solving within the [Jan App](https://jan.ai/). Based on our [**Lucy**](https://huggingface.co/Menlo/Lucy) model, Jan-v1 achieves improved performance through model scaling.
Jan-v1 leverages the newly released [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to deliver significantly enhanced reasoning capabilities and effective tool utilizatio. This architectural evolution is designed to deliver superior performance on complex agentic tasks, setting a new benchmark for accessible, high-performance AI.
Jan-v1 uses the [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks.
## Performance
### Question Answering (SimpleQA)
For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.2% accuracy.
For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/abEitIjvszFm7Z8mRHQz-.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/1_9-r0GpBACSsWoUVZ1Xg.png)
*The 91.2% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach.*
*The 91.1% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach.*
### Chat Benchmarks
These benchmarks evaluate the model's conversational and instructional capabilities.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/fX7BTAHaYdohIj4fbrIhM.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/f3bzNYRuA_iTFQIcvu6Rr.png)
## Quick Start
@@ -48,7 +48,7 @@ Jan-v1 is optimized for direct integration with the [Jan App](https://jan.ai/).
**Using vLLM:**
```bash
vllm serve Menlo/Jan-v1 \
vllm serve janhq/Jan-v1-4B \
--host 0.0.0.0 \
--port 1234 \
--enable-auto-tool-choice \
@@ -59,7 +59,9 @@ vllm serve Menlo/Jan-v1 \
```bash
llama-server --model jan-v1.gguf \
--host 0.0.0.0 \
--port 1234
--port 1234 \
--jinja \
--no-context-shift
```
### Recommended Parameters
@@ -75,9 +77,13 @@ max_tokens: 2048
## 🤝 Community & Support
- **Discussions**: [HuggingFace Community](https://huggingface.co/Menlo/Jan-v1/discussions) <!-- Update with your HF model ID -->
- **Discussions**: [HuggingFace Community](https://huggingface.co/janhq/Jan-v1-4B/discussions) <!-- Update with your HF model ID -->
- **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/)
## (*) Note
By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file [chat_template_raw.jinja](https://huggingface.co/janhq/Jan-v1-4B/blob/main/chat_template_raw.jinja).
## 📄 Citation
```bibtex
Updated Soon

View File

@@ -1,3 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:59842596e355cbdf25eeaffcd210ef179a1fd15206312072d431b616f055e1e8
size 21813275
oid sha256:bd99f33772b20b384a74efc5c386863a1f969651828f477b73b446583df5b0af
size 170956325