leaBroe commited on
Commit
ddc8dfc
·
verified ·
1 Parent(s): 4d09e3b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -2
README.md CHANGED
@@ -21,8 +21,54 @@ pip install -U adapters
21
  Now, the adapter can be loaded and activated like this:
22
 
23
  ```python
24
- from adapters import AutoAdapterModel
 
25
 
26
- model = EncoderDecoderModel.from_pretrained("leaBroe/Heavy2Light")
 
 
 
 
 
 
 
27
  adapter_name = model.load_adapter("leaBroe/Heavy2Light_adapter", set_active=True)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  ```
 
 
21
  Now, the adapter can be loaded and activated like this:
22
 
23
  ```python
24
+ from transformers import EncoderDecoderModel, AutoTokenizer, GenerationConfig
25
+ from adapters import init
26
 
27
+ model_path = "leaBroe/Heavy2Light"
28
+ subfolder_path = "heavy2light_final_checkpoint"
29
+
30
+ model = EncoderDecoderModel.from_pretrained(model_path)
31
+
32
+ tokenizer = AutoTokenizer.from_pretrained(model_path, subfolder=subfolder_path)
33
+
34
+ init(model)
35
  adapter_name = model.load_adapter("leaBroe/Heavy2Light_adapter", set_active=True)
36
+ model.set_active_adapters(adapter_name)
37
+ ```
38
+
39
+ then, the model can be used for inference:
40
+
41
+ ``` python
42
+ generation_config = GenerationConfig.from_pretrained(model_path)
43
+
44
+ # example input heavy sequence
45
+ heavy_seq = "QLQVQESGPGLVKPSETLSLTCTVSGASSSIKKYYWGWIRQSPGKGLEWIGSIYSSGSTQYNPALGSRVTLSVDTSQTQFSLRLTSVTAADTATYFCARQGADCTDGSCYLNDAFDVWGRGTVVTVSS"
46
+
47
+ inputs = tokenizer(
48
+ heavy_seq,
49
+ padding="max_length",
50
+ truncation=True,
51
+ max_length=250,
52
+ return_tensors="pt"
53
+ )
54
+
55
+ generated_seq = model.generate(
56
+ input_ids=inputs.input_ids,
57
+ attention_mask=inputs.attention_mask,
58
+ num_return_sequences=1,
59
+ output_scores=True,
60
+ return_dict_in_generate=True,
61
+ generation_config=generation_config,
62
+ bad_words_ids=[[4]],
63
+ do_sample=True,
64
+ temperature=1.0,
65
+ )
66
+
67
+ generated_text = tokenizer.decode(
68
+ generated_seq.sequences[0],
69
+ skip_special_tokens=True,
70
+ )
71
+
72
+ print("Generated light sequence:", generated_text)
73
  ```
74
+