| <! |
|
|
| Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with |
| the License. You may obtain a copy of the License at |
|
|
| http://www.apache.org/licenses/LICENSE-2.0 |
|
|
| Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on |
| an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the |
| specific language governing permissions and limitations under the License. |
| |
|
|
| |
|
|
| Hugging Face and Amazon introduced new [Hugging Face Deep Learning Containers (DLCs)](https://github.com/aws/deep-learning-containers/blob/master/available_images.md |
| make it easier than ever to train Hugging Face Transformer models in [Amazon SageMaker](https://aws.amazon.com/sagemaker/). |
|
|
| |
|
|
| |
|
|
|
|
| Before you can run your π€ Accelerate scripts on Amazon SageMaker you need to sign up for an AWS account. If you do not |
| have an AWS account yet learn more [here](https://docs.aws.amazon.com/sagemaker/latest/dg/gs-set-up.html). |
|
|
| After you have your AWS Account you need to install the `sagemaker` sdk for π€ Accelerate with: |
|
|
| ```bash |
| pip install "accelerate[sagemaker]" |
| ``` |
|
|
| π€ Accelerate currently uses the π€ DLCs, with `transformers`, `datasets` and `tokenizers` pre-installed. π€ |
| Accelerate is not in the DLC yet (will soon be added!) so to use it within Amazon SageMaker you need to create a |
| `requirements.txt` in the same directory where your training script is located and add it as dependency: |
|
|
| ``` |
| accelerate |
| ``` |
|
|
| You should also add any other dependencies you have to this `requirements.txt`. |
|
|
|
|
| |
|
|
| You can configure the launch configuration for Amazon SageMaker the same as you do for non SageMaker training jobs with |
| the π€ Accelerate CLI: |
|
|
| ```bash |
| accelerate config |
| |
| ``` |
|
|
| π€ Accelerate will go through a questionnaire about your Amazon SageMaker setup and create a config file you can edit. |
|
|
| <Tip> |
|
|
| π€ Accelerate is not saving any of your credentials. |
|
|
| </Tip> |
|
|
| |
|
|
| The training script is very similar to a training script you might run outside of SageMaker, but to save your model |
| after training you need to specify either `/opt/ml/model` or use `os.environ["SM_MODEL_DIR"]` as your save |
| directory. After training, artifacts in this directory are uploaded to S3: |
|
|
|
|
| ```diff |
| - torch.save('/opt/ml/model`) |
| + accelerator.save('/opt/ml/model') |
| ``` |
| |
| <Tip warning={true}> |
| |
| SageMaker doesnβt support argparse actions. If you want to use, for example, boolean hyperparameters, you need to |
| specify type as bool in your script and provide an explicit True or False value for this hyperparameter. [[REF]](https://sagemaker.readthedocs.io/en/stable/frameworks/pytorch/using_pytorch.html#prepare-a-pytorch-training-script). |
| |
| </Tip> |
| |
| ### Launch Training |
| |
| You can launch your training with π€ Accelerate CLI with: |
| |
| ``` |
| accelerate launch path_to_script.py --args_to_the_script |
| ``` |
| |
| This will launch your training script using your configuration. The only thing you have to do is provide all the |
| arguments needed by your training script as named arguments. |
| |
| **Examples** |
| |
| <Tip> |
| |
| If you run one of the example scripts, don't forget to add `accelerator.save('/opt/ml/model')` to it. |
|
|
| </Tip> |
|
|
| ```bash |
| accelerate launch ./examples/sagemaker_example.py |
| ``` |
|
|
| Outputs: |
|
|
| ``` |
| Configuring Amazon SageMaker environment |
| Converting Arguments to Hyperparameters |
| Creating Estimator |
| 2021-04-08 11:56:50 Starting - Starting the training job... |
| 2021-04-08 11:57:13 Starting - Launching requested ML instancesProfilerReport-1617883008: InProgress |
| ......... |
| 2021-04-08 11:58:54 Starting - Preparing the instances for training......... |
| 2021-04-08 12:00:24 Downloading - Downloading input data |
| 2021-04-08 12:00:24 Training - Downloading the training image.................. |
| 2021-04-08 12:03:39 Training - Training image download completed. Training in progress.. |
| ........ |
| epoch 0: {'accuracy': 0.7598039215686274, 'f1': 0.8178438661710037} |
| epoch 1: {'accuracy': 0.8357843137254902, 'f1': 0.882249560632689} |
| epoch 2: {'accuracy': 0.8406862745098039, 'f1': 0.8869565217391304} |
| ........ |
| 2021-04-08 12:05:40 Uploading - Uploading generated training model |
| 2021-04-08 12:05:40 Completed - Training job completed |
| Training seconds: 331 |
| Billable seconds: 331 |
| You can find your model data at: s3://your-bucket/accelerate-sagemaker-1-2021-04-08-11-56-47-108/output/model.tar.gz |
| ``` |
|
|
| |
|
|
| |
|
|
| Set up the accelerate config by running `accelerate config` and answer the SageMaker questions and set it up. |
| To use SageMaker DDP, select it when asked |
| `What is the distributed mode? ([0] No distributed training, [1] data parallelism):`. |
| Example config below: |
| ```yaml |
| base_job_name: accelerate-sagemaker-1 |
| compute_environment: AMAZON_SAGEMAKER |
| distributed_type: DATA_PARALLEL |
| ec2_instance_type: ml.p3.16xlarge |
| iam_role_name: xxxxx |
| image_uri: null |
| mixed_precision: fp16 |
| num_machines: 1 |
| profile: xxxxx |
| py_version: py38 |
| pytorch_version: 1.10.2 |
| region: us-east-1 |
| transformers_version: 4.17.0 |
| use_cpu: false |
| ``` |
|
|
| |
|
|
| *currently in development, will be supported soon.* |
|
|
| |
|
|
| π€ Accelerate currently uses the π€ DLCs, with `transformers`, `datasets` and `tokenizers` pre-installed. If you |
| want to use different/other Python packages you can do this by adding them to the `requirements.txt`. These packages |
| will be installed before your training script is started. |
|
|
| |
|
|
| *undecided if feature is needed. Contact us if you would like this feature.* |
|
|
| |
|
|
| *undecided if feature is needed. Contact us if you would like this feature.* |
|
|