Update README.md
Browse files
README.md
CHANGED
|
@@ -24,7 +24,7 @@ pipeline_tag: image-text-to-text
|
|
| 24 |
## Introduction
|
| 25 |
|
| 26 |
We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
|
| 27 |
-
Built upon
|
| 28 |
|
| 29 |
## Features
|
| 30 |
|
|
@@ -32,7 +32,7 @@ Built upon a 8B dense language model (Qwen3) and a 400M Vision encoder (InternVi
|
|
| 32 |
|
| 33 |
- Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
|
| 34 |
|
| 35 |
-
- Dynamic tokenizer enables native understanding of molecular formulas
|
| 36 |
|
| 37 |
## Performance
|
| 38 |
|
|
@@ -139,7 +139,7 @@ print(decoded_output)
|
|
| 139 |
|
| 140 |
#### Video input
|
| 141 |
|
| 142 |
-
Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please
|
| 143 |
|
| 144 |
```python
|
| 145 |
from transformers import AutoProcessor, AutoModelForCausalLM
|
|
@@ -385,7 +385,7 @@ print(response.choices[0].message.content)
|
|
| 385 |
|
| 386 |
### Switching Between Thinking and Non-Thinking Modes
|
| 387 |
|
| 388 |
-
Intern-S1 enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
|
| 389 |
|
| 390 |
```python
|
| 391 |
text = tokenizer.apply_chat_template(
|
|
@@ -396,7 +396,7 @@ text = tokenizer.apply_chat_template(
|
|
| 396 |
)
|
| 397 |
```
|
| 398 |
|
| 399 |
-
With LMDeploy serving Intern-S1 models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
|
| 400 |
|
| 401 |
```python
|
| 402 |
from openai import OpenAI
|
|
|
|
| 24 |
## Introduction
|
| 25 |
|
| 26 |
We introduce **Intern-S1-mini**, a lightweight open-source multimodal reasoning model based on the same techniques as **[Intern-S1](https://huggingface.co/internlm/Intern-S1)**.
|
| 27 |
+
Built upon an 8B dense language model (Qwen3) and a 400M Vision encoder (InternViT), Intern-S1-mini has been further pretrained on **5 trillion tokens** of multimodal data, including over **2.5 trillion scientific-domain tokens**. This enables the model to retain strong general capabilities while excelling in specialized scientific domains such as **interpreting chemical structures, understanding protein sequences, and planning compound synthesis routes**, making Intern-S1-mini to be a capable research assistant for real-world scientific applications.
|
| 28 |
|
| 29 |
## Features
|
| 30 |
|
|
|
|
| 32 |
|
| 33 |
- Continuously pretrained on a massive 5T token dataset, with over 50% specialized scientific data, embedding deep domain expertise.
|
| 34 |
|
| 35 |
+
- Dynamic tokenizer enables native understanding of molecular formulas and protein sequences.
|
| 36 |
|
| 37 |
## Performance
|
| 38 |
|
|
|
|
| 139 |
|
| 140 |
#### Video input
|
| 141 |
|
| 142 |
+
Please ensure that the decord video decoding library is installed via `pip install decord`. To avoid OOM, please install flash_attention and use at least 2 GPUS.
|
| 143 |
|
| 144 |
```python
|
| 145 |
from transformers import AutoProcessor, AutoModelForCausalLM
|
|
|
|
| 385 |
|
| 386 |
### Switching Between Thinking and Non-Thinking Modes
|
| 387 |
|
| 388 |
+
Intern-S1-mini enables thinking mode by default, enhancing the model's reasoning capabilities to generate higher-quality responses. This feature can be disabled by setting `enable_thinking=False` in `tokenizer.apply_chat_template`
|
| 389 |
|
| 390 |
```python
|
| 391 |
text = tokenizer.apply_chat_template(
|
|
|
|
| 396 |
)
|
| 397 |
```
|
| 398 |
|
| 399 |
+
With LMDeploy serving Intern-S1-mini models, you can dynamically control the thinking mode by adjusting the `enable_thinking` parameter in your requests.
|
| 400 |
|
| 401 |
```python
|
| 402 |
from openai import OpenAI
|