RangiLyu commited on
Commit
c2e8223
·
verified ·
1 Parent(s): 12e6d31

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -24,7 +24,7 @@ pipeline_tag: image-text-to-text
24
  ## Introduction
25
 
26
  We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
27
- Built upon a 8B dense language model (Qwen3) and a 400M Vision encoder (InternViT), Intern-S1-mini has been further pretrained on **5 trillion tokens** of multimodal data, including over **2.5 trillion scientific-domain tokens**. This enables the model to retain strong general capabilities while excelling in specialized scientific domains such as **interpreting chemical structures, understanding protein sequences, and planning compound synthesis routes**, making Intern-S1-mini to be a capable research assistant for real-world scientific applications.
28
 
29
  ## Features
30
 
@@ -32,7 +32,7 @@ Built upon a 8B dense language model (Qwen3) and a 400M Vision encoder (InternVi
32
 
33
  - Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
34
 
35
- - Dynamic tokenizer enables native understanding of molecular formulas, protein sequences, and seismic signals.
36
 
37
  ## Performance
38
 
@@ -139,7 +139,7 @@ print(decoded_output)
139
 
140
  #### Video input
141
 
142
- Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please at least use 2 GPUS.
143
 
144
  ```python
145
  from transformers import AutoProcessor, AutoModelForCausalLM
@@ -385,7 +385,7 @@ print(response.choices[0].message.content)
385
 
386
  ### Switching Between Thinking and Non-Thinking Modes
387
 
388
- Intern-S1 enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
389
 
390
  ```python
391
  text = tokenizer.apply_chat_template(
@@ -396,7 +396,7 @@ text = tokenizer.apply_chat_template(
396
  )
397
  ```
398
 
399
- With LMDeploy serving Intern-S1 models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
400
 
401
  ```python
402
  from openai import OpenAI
 
24
  ## Introduction
25
 
26
  We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
27
+ Built upon an 8B dense language model (Qwen3) and a 400M Vision encoder (InternViT), Intern-S1-mini has been further pretrained on **5 trillion tokens** of multimodal data, including over **2.5 trillion scientific-domain tokens**. This enables the model to retain strong general capabilities while excelling in specialized scientific domains such as **interpreting chemical structures, understanding protein sequences, and planning compound synthesis routes**, making Intern-S1-mini to be a capable research assistant for real-world scientific applications.
28
 
29
  ## Features
30
 
 
32
 
33
  - Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
34
 
35
+ - Dynamic tokenizer enables native understanding of molecular formulas and protein sequences.
36
 
37
  ## Performance
38
 
 
139
 
140
  #### Video input
141
 
142
+ Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please install flash_attention and use at least 2 GPUS.
143
 
144
  ```python
145
  from transformers import AutoProcessor, AutoModelForCausalLM
 
385
 
386
  ### Switching Between Thinking and Non-Thinking Modes
387
 
388
+ Intern-S1-mini enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
389
 
390
  ```python
391
  text = tokenizer.apply_chat_template(
 
396
  )
397
  ```
398
 
399
+ With LMDeploy serving Intern-S1-mini models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
400
 
401
  ```python
402
  from openai import OpenAI