Update README.md
Browse files
README.md
CHANGED
@@ -5,4 +5,84 @@ language:
|
|
5 |
base_model:
|
6 |
- Qwen/Qwen3-4B-Thinking-2507
|
7 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
---
|
|
|
5 |
base_model:
|
6 |
- Qwen/Qwen3-4B-Thinking-2507
|
7 |
pipeline_tag: text-generation
|
8 |
+
---
|
9 |
+
# Jan-v1: The Inaugural Model of the Jan Family – Redefining Agentic Reasoning
|
10 |
+
|
11 |
+
[](https://github.com/menloresearch/deep-research)
|
12 |
+
[](https://opensource.org/licenses/Apache-2.0)
|
13 |
+
[](https://jan.ai/) # Adding a badge for Jan App
|
14 |
+
|
15 |
+
<div align="center">
|
16 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/PA6JCiYLPJX_WFO42ClTd.jpeg" width="300" alt="Jan-v1-Demo-Image"> <!-- Placeholder or replace with actual Jan-v1 specific image/GIF -->
|
17 |
+
</div>
|
18 |
+
|
19 |
+
**Authors:** [Alan Dao](https://scholar.google.com/citations?user=eGWws2UAAAAJ&hl=en), [Bach Vu Dinh](https://scholar.google.com/citations?user=7Lr6hdoAAAAJ&hl=vi), [Alex Nguyen](https://github.com/nguyenhoangthuan99), [Norapat Buppodom](https://scholar.google.com/citations?user=utfEThsAAAAJ&hl=th&authuser=1) # Assuming same authors, adjust if needed.
|
20 |
+
|
21 |
+
<!-- Optional: If you have a GIF for Jan-v1, include it here like Lucy's. -->
|
22 |
+
<!--  -->
|
23 |
+
|
24 |
+
## Overview
|
25 |
+
|
26 |
+
Introducing **Jan-v1**, the foundational model in the **Jan Family** – a new lineage of highly capable language models developed to power the next generation of intelligent agents within the [Jan App](https://jan.ai/) ecosystem. Building on the innovative agentic capabilities of our earlier **Lucy** model, Jan-v1 represents a significant leap forward through strategic model scaling.
|
27 |
+
|
28 |
+
By leveraging a larger **Qwen3-4B** base, Jan-v1 demonstrates profoundly enhanced 'thinking' and reasoning capabilities. This architectural evolution is designed to deliver superior performance on complex agentic tasks, setting a new benchmark for accessible, high-performance AI.
|
29 |
+
|
30 |
+
## What Jan-v1 Excels At
|
31 |
+
|
32 |
+
- **🧠 Enhanced Agentic Reasoning**: With its larger parameter count, Jan-v1 excels at deeper reasoning, complex problem-solving, and sophisticated multi-step agentic planning.
|
33 |
+
- **🎯 Superior Question Answering**: Achieves an impressive **91.2% accuracy on SimpleQA**, significantly advancing performance for factoid question answering.
|
34 |
+
- **🔍 Advanced Agentic Web Search**: Inherits and refines Lucy's strong capabilities for agentic web search and lightweight browsing via MCP-enabled tools.
|
35 |
+
- **📱 Optimized for Jan App**: Specifically engineered to provide unique and highly optimized support for the Jan App, ensuring seamless integration and superior user experience.
|
36 |
+
|
37 |
+
## Evaluation
|
38 |
+
|
39 |
+
Jan-v1's strategic scaling has resulted in a notable performance uplift, particularly evident in its "thinking" and reasoning prowess. Following the established MCP benchmark methodology, Jan-v1 sets a new standard for models in its class.
|
40 |
+
|
41 |
+
| Model | SimpleQA Accuracy |
|
42 |
+
| :---------------------------------- | :---------------- |
|
43 |
+
| **Jan-v1 (Qwen3-4B)** | **91.2%** |
|
44 |
+
| Lucy (Qwen3-1.7B) | [Lucy's Score] | <!-- Insert Lucy's actual SimpleQA score from its README here for direct comparison -->
|
45 |
+
| DeepSeek-v3 (Comparison from Lucy) | [DeepSeek's Score]| <!-- Insert DeepSeek's score from Lucy's README here -->
|
46 |
+
|
47 |
+
The **91.2% accuracy on SimpleQA** underscores Jan-v1's advanced ability to precisely retrieve and synthesize information, showcasing the effectiveness of our model scaling approach for agentic intelligence.
|
48 |
+
|
49 |
+
## 🖥️ How to Run Locally
|
50 |
+
|
51 |
+
Jan-v1 is designed for flexible deployment, compatible with various inference engines including vLLM, llama.cpp, and local applications like Jan and LMStudio. Its integration with search APIs and web browsing tools is facilitated through the MCP (Mobile-Cloud Protocol).
|
52 |
+
|
53 |
+
### Deployment
|
54 |
+
|
55 |
+
Deploy using VLLM:
|
56 |
+
```bash
|
57 |
+
vllm serve Menlo/Jan-v1 \ # Update with your HF model ID (e.g., Menlo/Jan-v1)
|
58 |
+
--host 0.0.0.0 \
|
59 |
+
--port 1234 \
|
60 |
+
--enable-auto-tool-choice \
|
61 |
+
--tool-call-parser hermes
|
62 |
+
```
|
63 |
+
|
64 |
+
Or `llama-server` from `llama.cpp`:
|
65 |
+
```bash
|
66 |
+
llama-server ...
|
67 |
+
```
|
68 |
+
|
69 |
+
### Recommended Sampling Parameters
|
70 |
+
|
71 |
+
```yaml
|
72 |
+
Temperature: 0.7
|
73 |
+
Top-p: 0.9
|
74 |
+
Top-k: 20
|
75 |
+
Min-p: 0.0
|
76 |
+
```
|
77 |
+
|
78 |
+
## 🤝 Community & Support
|
79 |
+
|
80 |
+
- **Discussions**: [HuggingFace Community](https://huggingface.co/Menlo/Jan-v1/discussions) <!-- Update with your HF model ID -->
|
81 |
+
- **Jan App**: Learn more about the Jan App at [jan.ai](https://jan.ai/)
|
82 |
+
|
83 |
+
## 📄 Citation
|
84 |
+
```bibtex
|
85 |
+
Updated Soon
|
86 |
+
```
|
87 |
+
**Paper **: *Jan-v1
|
88 |
---
|