4llengodlike commited on
Commit
8cc6b8b
·
verified ·
1 Parent(s): 0542863

Upload 2 files

Browse files
Files changed (2) hide show
  1. config.json +44 -0
  2. handler.py +27 -0
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "llama",
3
+ "hidden_size": 4096,
4
+ "num_attention_heads": 16,
5
+ "num_hidden_layers": 32,
6
+ "intermediate_size": 11008,
7
+ "hidden_act": "gelu",
8
+ "initializer_range": 0.02,
9
+ "layer_norm_eps": 1e-5,
10
+ "max_position_embeddings": 2048,
11
+ "vocab_size": 32000,
12
+ "model_name": "LlamaForSequenceClassification",
13
+ "pipeline_tag": "text-generation",
14
+ "peft_config": {
15
+ "r": 16,
16
+ "target_modules": [
17
+ "q_proj", "k_proj", "v_proj", "o_proj",
18
+ "gate_proj", "up_proj", "down_proj",
19
+ "embed_tokens", "lm_head"
20
+ ],
21
+ "lora_alpha": 16,
22
+ "lora_dropout": 0,
23
+ "bias": "none",
24
+ "use_gradient_checkpointing": "unsloth",
25
+ "random_state": 3407,
26
+ "use_rslora": false,
27
+ "loftq_config": null
28
+ },
29
+ "training_args": {
30
+ "per_device_train_batch_size": 2,
31
+ "gradient_accumulation_steps": 4,
32
+ "warmup_steps": 5,
33
+ "max_steps": 60,
34
+ "learning_rate": 2e-4,
35
+ "fp16": true,
36
+ "bf16": false,
37
+ "logging_steps": 1,
38
+ "optim": "adamw_8bit",
39
+ "weight_decay": 0.01,
40
+ "lr_scheduler_type": "linear",
41
+ "seed": 3407,
42
+ "output_dir": "outputs"
43
+ }
44
+ }
handler.py ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer
2
+ import torch
3
+
4
+ class ModelHandler:
5
+ def __init__(self):
6
+ self.model = None
7
+ self.tokenizer = None
8
+
9
+ def load_model(self):
10
+ # 加载模型和分词器
11
+ self.model = AutoModelForCausalLM.from_pretrained("your-model-path")
12
+ self.tokenizer = AutoTokenizer.from_pretrained("your-model-path")
13
+
14
+ def predict(self, inputs):
15
+ # 将输入转换为模型可以处理的格式
16
+ inputs = self.tokenizer(inputs, return_tensors="pt")
17
+ with torch.no_grad():
18
+ outputs = self.model(**inputs)
19
+ return outputs
20
+
21
+ handler = ModelHandler()
22
+ handler.load_model()
23
+
24
+ def handler(event, context):
25
+ inputs = event["data"]
26
+ outputs = handler.predict(inputs)
27
+ return outputs