import sagemaker import boto3 from sagemaker.huggingface import HuggingFaceModel try: role = sagemaker.get_execution_role() except ValueError: iam = boto3.client('iam') role = iam.get_role(RoleName='sagemaker_execution_role')['Role']['Arn'] # Hub Model configuration. https://huggingface.co/models hub = { 'HF_MODEL_ID':'Qwen/Qwen2.5-Omni-7B', 'HF_TASK':'any-to-any' } # create Hugging Face Model Class huggingface_model = HuggingFaceModel( transformers_version='4.49.0', pytorch_version='2.6.0', py_version='py312', env=hub, role=role, ) # deploy model to SageMaker Inference predictor = huggingface_model.deploy( initial_instance_count=1, # number of instances instance_type='ml.m5.xlarge' # ec2 instance type )

#31
by Moises1m - opened
Files changed (1) hide show
  1. .gitattributes +31 -0
.gitattributes CHANGED
@@ -30,3 +30,34 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
30
  *.zstandard filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
  model.safetensors filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  *.zstandard filter=lfs diff=lfs merge=lfs -text
31
  *tfevents* filter=lfs diff=lfs merge=lfs -text
32
  model.safetensors filter=lfs diff=lfs merge=lfs -text
33
+ import sagemaker
34
+ import boto3
35
+ from sagemaker.huggingface import HuggingFaceModel
36
+
37
+ try:
38
+ role = sagemaker.get_execution_role()
39
+ except ValueError:
40
+ iam = boto3.client('iam')
41
+ role = iam.get_role(RoleName='sagemaker_execution_role')['Role']['Arn']
42
+
43
+ # Hub Model configuration. https://huggingface.co/models
44
+ hub = {
45
+ 'HF_MODEL_ID':'Qwen/Qwen2.5-Omni-7B',
46
+ 'HF_TASK':'any-to-any'
47
+ }
48
+
49
+ # create Hugging Face Model Class
50
+ huggingface_model = HuggingFaceModel(
51
+ transformers_version='4.49.0',
52
+ pytorch_version='2.6.0',
53
+ py_version='py312',
54
+ env=hub,
55
+ role=role,
56
+ )
57
+
58
+ # deploy model to SageMaker Inference
59
+ predictor = huggingface_model.deploy(
60
+ initial_instance_count=1, # number of instances
61
+ instance_type='ml.m5.xlarge' # ec2 instance type
62
+ )
63
+