Update README.md
Browse files
README.md
CHANGED
@@ -24,7 +24,7 @@ pipeline_tag: image-text-to-text
|
|
24 |
## Introduction
|
25 |
|
26 |
We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
|
27 |
-
Built upon
|
28 |
|
29 |
## Features
|
30 |
|
@@ -32,7 +32,7 @@ Built upon a 8B dense language model (Qwen3) and a 400M Vision encoder (InternVi
|
|
32 |
|
33 |
- Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
|
34 |
|
35 |
-
- Dynamic tokenizer enables native understanding of molecular formulas
|
36 |
|
37 |
## Performance
|
38 |
|
@@ -139,7 +139,7 @@ print(decoded_output)
|
|
139 |
|
140 |
#### Video input
|
141 |
|
142 |
-
Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please at least
|
143 |
|
144 |
```python
|
145 |
from transformers import AutoProcessor, AutoModelForCausalLM
|
@@ -385,7 +385,7 @@ print(response.choices[0].message.content)
|
|
385 |
|
386 |
### Switching Between Thinking and Non-Thinking Modes
|
387 |
|
388 |
-
Intern-S1 enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
|
389 |
|
390 |
```python
|
391 |
text = tokenizer.apply_chat_template(
|
@@ -396,7 +396,7 @@ text = tokenizer.apply_chat_template(
|
|
396 |
)
|
397 |
```
|
398 |
|
399 |
-
With LMDeploy serving Intern-S1 models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
|
400 |
|
401 |
```python
|
402 |
from openai import OpenAI
|
|
|
24 |
## Introduction
|
25 |
|
26 |
We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
|
27 |
+
Built upon an 8B dense language model (Qwen3) and a 400M Vision encoder (InternViT), Intern-S1-mini has been further pretrained on **5 trillion tokens** of multimodal data, including over **2.5 trillion scientific-domain tokens**. This enables the model to retain strong general capabilities while excelling in specialized scientific domains such as **interpreting chemical structures, understanding protein sequences, and planning compound synthesis routes**, making Intern-S1-mini to be a capable research assistant for real-world scientific applications.
|
28 |
|
29 |
## Features
|
30 |
|
|
|
32 |
|
33 |
- Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
|
34 |
|
35 |
+
- Dynamic tokenizer enables native understanding of molecular formulas and protein sequences.
|
36 |
|
37 |
## Performance
|
38 |
|
|
|
139 |
|
140 |
#### Video input
|
141 |
|
142 |
+
Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please install flash_attention and use at least 2 GPUS.
|
143 |
|
144 |
```python
|
145 |
from transformers import AutoProcessor, AutoModelForCausalLM
|
|
|
385 |
|
386 |
### Switching Between Thinking and Non-Thinking Modes
|
387 |
|
388 |
+
Intern-S1-mini enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
|
389 |
|
390 |
```python
|
391 |
text = tokenizer.apply_chat_template(
|
|
|
396 |
)
|
397 |
```
|
398 |
|
399 |
+
With LMDeploy serving Intern-S1-mini models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
|
400 |
|
401 |
```python
|
402 |
from openai import OpenAI
|