d-s-b commited on
Commit
20931d0
·
verified ·
1 Parent(s): a7aad0e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -6
README.md CHANGED
@@ -3,10 +3,12 @@ base_model: google/gemma-3-270m-it
3
  library_name: transformers
4
  model_name: Router
5
  tags:
6
- - generated_from_trainer
7
  - trl
8
  - sft
 
9
  licence: license
 
 
10
  ---
11
 
12
  # Model Card for Router
@@ -19,10 +21,11 @@ It has been trained using [TRL](https://github.com/huggingface/trl).
19
  ```python
20
  from transformers import pipeline
21
 
22
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="d-s-b/Router", device="cuda")
24
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
- print(output["generated_text"])
 
26
  ```
27
 
28
  ## Training procedure
@@ -30,7 +33,7 @@ print(output["generated_text"])
30
 
31
 
32
 
33
- This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
@@ -55,4 +58,13 @@ Cite TRL as:
55
  publisher = {GitHub},
56
  howpublished = {\url{https://github.com/huggingface/trl}}
57
  }
 
 
 
 
 
 
 
 
 
58
  ```
 
3
  library_name: transformers
4
  model_name: Router
5
  tags:
 
6
  - trl
7
  - sft
8
+ - gemma3
9
  licence: license
10
+ datasets:
11
+ - d-s-b/synthetic-reasoning-dataset
12
  ---
13
 
14
  # Model Card for Router
 
21
  ```python
22
  from transformers import pipeline
23
 
24
+ pipe = pipeline("text-generation", model="d-s-b/Router")
25
+ messages = [
26
+ {"role": "user", "content": "what is capital of india"}
27
+ ]
28
+ pipe(messages)
29
  ```
30
 
31
  ## Training procedure
 
33
 
34
 
35
 
36
+ This model was trained with SFT
37
 
38
  ### Framework versions
39
 
 
58
  publisher = {GitHub},
59
  howpublished = {\url{https://github.com/huggingface/trl}}
60
  }
61
+ @article{gemma_2025,
62
+ title={Gemma 3},
63
+ url={https://arxiv.org/abs/2503.19786},
64
+ publisher={Google DeepMind},
65
+ author={Gemma Team},
66
+ year={2025}
67
+ }
68
+
69
+
70
  ```