Update README.md
Browse files
README.md
CHANGED
@@ -13,8 +13,13 @@ datasets:
|
|
13 |
|
14 |
# Model Card for Router
|
15 |
|
16 |
-
This model is
|
17 |
-
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
## Quick start
|
20 |
|
@@ -28,12 +33,24 @@ messages = [
|
|
28 |
pipe(messages)
|
29 |
```
|
30 |
|
31 |
-
## Training
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
-
|
34 |
|
|
|
35 |
|
36 |
-
|
37 |
|
38 |
### Framework versions
|
39 |
|
|
|
13 |
|
14 |
# Model Card for Router
|
15 |
|
16 |
+
This model is fine-tuned to serve as a router for reasoning tasks, classifying input queries into one of three categories:
|
17 |
+
|
18 |
+
no_reasoning – Direct factual lookup or simple recall (e.g., "What is the capital of France?")
|
19 |
+
|
20 |
+
low_reasoning – Requires light reasoning such as simple arithmetic, comparisons, or single logical steps (e.g., "If John has 5 apples and eats 2, how many are left?")
|
21 |
+
|
22 |
+
high_reasoning – Requires multi-step reasoning, deep logical chains, or complex problem-solving (e.g., "Prove that the sum of two even numbers is always even").
|
23 |
|
24 |
## Quick start
|
25 |
|
|
|
33 |
pipe(messages)
|
34 |
```
|
35 |
|
36 |
+
## Training Details
|
37 |
+
|
38 |
+
Method: Supervised fine-tuning with SFTTrainer
|
39 |
+
|
40 |
+
Objective: Multi-class classification with labels (no_reasoning, low_reasoning, high_reasoning)
|
41 |
+
|
42 |
+
Dataset: Custom dataset of queries annotated with reasoning levels.
|
43 |
+
|
44 |
+
|
45 |
+
|
46 |
+
|
47 |
+
## Limitations & Bias
|
48 |
|
49 |
+
May misclassify borderline queries (e.g., between low_reasoning and high_reasoning).
|
50 |
|
51 |
+
Performance depends on the diversity of training data.
|
52 |
|
53 |
+
Inherits any biases from the base Gemma 3 270M model.
|
54 |
|
55 |
### Framework versions
|
56 |
|