lhallee commited on
Commit
238b112
·
verified ·
1 Parent(s): 5a656ba

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +46 -46
README.md CHANGED
@@ -1,46 +1,46 @@
1
- ---
2
- library_name: transformers
3
- tags: []
4
- ---
5
-
6
- # NOTE
7
- The GitHub with the implementation and requirements can be found [here](https://github.com/Synthyra/FastPLMs.git).
8
-
9
- # DPLM2
10
- Synthyra DPLM2 checkpoints are HuggingFace AutoModel compatible and include FastPLMs embedding helpers.
11
-
12
- ## Supported models
13
- ```python
14
- model_dict = {
15
- "Synthyra/DPLM2-150M": "airkingbd/dplm2_150m",
16
- "Synthyra/DPLM2-650M": "airkingbd/dplm2_650m",
17
- "Synthyra/DPLM2-3B": "airkingbd/dplm2_3b",
18
- }
19
- ```
20
-
21
- ## Use with transformers
22
- ```python
23
- import torch
24
- from transformers import AutoModel, AutoModelForMaskedLM
25
-
26
- model_path = "Synthyra/DPLM2-150M"
27
- model = AutoModel.from_pretrained(model_path, trust_remote_code=True, dtype=torch.float16).eval()
28
- tokenizer = model.tokenizer
29
-
30
- batch = tokenizer(["MPRTEIN", "MSEQWENCE"], padding=True, return_tensors="pt")
31
- with torch.no_grad():
32
- hidden = model(**batch).last_hidden_state
33
-
34
- mlm = AutoModelForMaskedLM.from_pretrained(model_path, trust_remote_code=True, dtype=torch.float16).eval()
35
- with torch.no_grad():
36
- logits = mlm(**batch).logits
37
- ```
38
-
39
- ## DPLM2 modality types
40
- DPLM2 infers `type_ids` automatically from `input_ids` and `attention_mask` when they are not provided.
41
-
42
- ## Attention backend
43
- `sdpa` is the default backend. Flex Attention is available by setting `config.attn_backend = "flex"` before loading.
44
-
45
- ## Embed datasets
46
- All DPLM2 models inherit `EmbeddingMixin`, so you can call `model.embed_dataset(...)` directly.
 
1
+ ---
2
+ library_name: transformers
3
+ tags: []
4
+ ---
5
+
6
+ # NOTE
7
+ The GitHub with the implementation and requirements can be found [here](https://github.com/Synthyra/FastPLMs.git).
8
+
9
+ # DPLM2
10
+ Synthyra DPLM2 checkpoints are HuggingFace AutoModel compatible and include FastPLMs embedding helpers.
11
+
12
+ ## Supported models
13
+ ```python
14
+ model_dict = {
15
+ "Synthyra/DPLM2-150M": "airkingbd/dplm2_150m",
16
+ "Synthyra/DPLM2-650M": "airkingbd/dplm2_650m",
17
+ "Synthyra/DPLM2-3B": "airkingbd/dplm2_3b",
18
+ }
19
+ ```
20
+
21
+ ## Use with transformers
22
+ ```python
23
+ import torch
24
+ from transformers import AutoModel, AutoModelForMaskedLM
25
+
26
+ model_path = "Synthyra/DPLM2-150M"
27
+ model = AutoModel.from_pretrained(model_path, trust_remote_code=True, dtype=torch.float16).eval()
28
+ tokenizer = model.tokenizer
29
+
30
+ batch = tokenizer(["MPRTEIN", "MSEQWENCE"], padding=True, return_tensors="pt")
31
+ with torch.no_grad():
32
+ hidden = model(**batch).last_hidden_state
33
+
34
+ mlm = AutoModelForMaskedLM.from_pretrained(model_path, trust_remote_code=True, dtype=torch.float16).eval()
35
+ with torch.no_grad():
36
+ logits = mlm(**batch).logits
37
+ ```
38
+
39
+ ## DPLM2 modality types
40
+ DPLM2 infers `type_ids` automatically from `input_ids` and `attention_mask` when they are not provided.
41
+
42
+ ## Attention backend
43
+ `sdpa` is the default backend. Flex Attention is available by setting `config.attn_backend = "flex"` before loading.
44
+
45
+ ## Embed datasets
46
+ All DPLM2 models inherit `EmbeddingMixin`, so you can call `model.embed_dataset(...)` directly.