lucyknada commited on
Commit
8e8797c
·
verified ·
1 Parent(s): 3c123f9

Upload ./README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +92 -0
README.md ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen3-4B-Thinking-2507
7
+ pipeline_tag: text-generation
8
+ ---
9
+ ### exl3 quant
10
+ ---
11
+ ### check revisions for quants
12
+ ---
13
+
14
+ # Jan-v1: Advanced Agentic Language Model
15
+
16
+ [![GitHub](https://img.shields.io/badge/GitHub-Repository-blue?logo=github)](https://github.com/menloresearch/deep-research)
17
+ [![License](https://img.shields.io/badge/License-Apache%202.0-yellow)](https://opensource.org/licenses/Apache-2.0)
18
+ [![Jan App](https://img.shields.io/badge/Powered%20by-Jan%20App-purple?style=flat&logo=android)](https://jan.ai/)
19
+
20
+ <!-- Optional: If you have a GIF for Jan-v1, include it here like Lucy's. -->
21
+ <!-- ![image/gif](jan_v1_demo.gif) -->
22
+
23
+ ## Overview
24
+ **Jan-v1** is the first release in the **Jan Family**, designed for agentic reasoning and problem-solving within the [Jan App](https://jan.ai/). Based on our [**Lucy**](https://huggingface.co/Menlo/Lucy) model, Jan-v1 achieves improved performance through model scaling.
25
+
26
+ Jan-v1 uses the [Qwen3-4B-thinking](https://huggingface.co/Qwen/Qwen3-4B-Thinking-2507) model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks.
27
+
28
+ ## Performance
29
+
30
+ ### Question Answering (SimpleQA)
31
+ For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy.
32
+
33
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/6CaETynCW18MXgDrbp_N9.png)
34
+
35
+ *The 91.1% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach.*
36
+
37
+ ### Chat Benchmarks
38
+
39
+ These benchmarks evaluate the model's conversational and instructional capabilities.
40
+
41
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/f3bzNYRuA_iTFQIcvu6Rr.png)
42
+
43
+ ## Quick Start
44
+
45
+ ### Integration with Jan App
46
+
47
+ Jan-v1 is optimized for direct integration with the [Jan App](https://jan.ai/). Simply select the model from the Jan App interface for immediate access to its full capabilities.
48
+
49
+ ![image/gif](demo.gif)
50
+
51
+ ### Local Deployment
52
+
53
+ **Using vLLM:**
54
+ ```bash
55
+ vllm serve janhq/Jan-v1-4B \
56
+ --host 0.0.0.0 \
57
+ --port 1234 \
58
+ --enable-auto-tool-choice \
59
+ --tool-call-parser hermes
60
+
61
+ ```
62
+
63
+ **Using llama.cpp:**
64
+ ```bash
65
+ llama-server --model Jan-v1-4B-Q4_K_M.gguf \
66
+ --host 0.0.0.0 \
67
+ --port 1234 \
68
+ --jinja \
69
+ --no-context-shift
70
+ ```
71
+
72
+ ### Recommended Parameters
73
+
74
+ ```yaml
75
+ temperature: 0.6
76
+ top_p: 0.95
77
+ top_k: 20
78
+ min_p: 0.0
79
+ max_tokens: 2048
80
+ ```
81
+
82
+
83
+ ## 🤝 Community & Support
84
+
85
+ - **Discussions**: [HuggingFace Community](https://huggingface.co/janhq/Jan-v1-4B/discussions)
86
+ - **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/)
87
+
88
+ ## 📄 Citation
89
+ ```bibtex
90
+ Updated Soon
91
+ ```
92
+ ---