File size: 2,057 Bytes
f49e2e5
 
9ca9af4
 
 
 
f49e2e5
 
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
6330e4a
dc77eb0
 
f49e2e5
4ae0ef5
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
 
 
f49e2e5
9ca9af4
dca4e8a
 
 
f49e2e5
dca4e8a
 
 
 
f49e2e5
9ca9af4
 
 
0270c7e
9ca9af4
 
 
 
 
 
 
 
 
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
f49e2e5
9ca9af4
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
---
library_name: transformers
license: mit
language:
- ja
- en
---

# Stockmark-2-100B-Instruct-beta

![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/607ef1c3e758c3c5a2959eab/AbyPvKu-FBY6RDYhGi1KX.jpeg)

## Model description

**Stockmark-2-100B-Instruct-beta** is a 100-billion-parameter large language model built from scratch, with a particular focus on Japanese. It was pre-trained on approximately 1.5 trillion tokens of data, consisting of 60% English, 30% Japanese, and 10% code. Following pretraining, the model underwent post-training with synthetic data in Japanese to enhance its ability to follow instructions. This synthetic data was generated using Qwen2.5-32B-Instruct.

As a beta release, Stockmark-2-100b-Instruct-beta is still undergoing improvements and evaluations. Feedback and insights from users will help refine future versions.

See [our blog](https://stockmark-tech.hatenablog.com/entry/2025/03/06/114203) for the detail.

This project is supported by [GENIAC](https://www.meti.go.jp/policy/mono_info_service/geniac/index.html).

## How to use

```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("stockmark/Stockmark-2-100B-Instruct-beta")
model = AutoModelForCausalLM.from_pretrained(
    "stockmark/Stockmark-2-100B-Instruct-beta", device_map="auto", torch_dtype=torch.bfloat16
)

instruction = "自然言語処理とは?"
input_ids = tokenizer.apply_chat_template(
    [{"role": "user", "content": instruction}], add_generation_prompt=True, return_tensors="pt"
).to(model.device)

with torch.inference_mode():
    tokens = model.generate(
        input_ids,
        max_new_tokens = 512,
        do_sample = True,
        temperature = 0.7,
        top_p = 0.95,
        repetition_penalty = 1.05
    )
    
output = tokenizer.decode(tokens[0], skip_special_tokens=True)
print(output)
```

## License

[MIT](https://opensource.org/licenses/MIT)

## Developed by

[Stockmark Inc.](https://stockmark.co.jp/)

## Author

Takahiro Omi