Amazon SageMaker HyperPod recipes help customers get started with training and fine-tuning popular publicly available foundation models in just minutes, with state-of-the-art performance. The recipes provide a pre-configured training stack that is tested and validated on Amazon SageMaker.
Please see Amazon SageMaker HyperPod recipes for documentation.
The recipes support Amazon SageMaker HyperPod (with Slurm or Amazon EKS for workload orchestration) and Amazon SageMaker training jobs.
Amazon SageMaker HyperPod recipes include built-in support for:
- Model parallelism - tensor parallelism and context parallel
- Automated distributed checkpointing
- Distributed optimizer
- Accelerators: NVIDIA H100 (ml.p5), NVIDIA A100 (ml.p4), and AWS Trainium (ml.trn1)
- Fine-tuning: Full, QLoRA, LoRA
- AWS Instances: ml.p5.48xlarge, ml.p4d.24xlarge, and ml.trn1.32xlarge instance families
- Supported Models: Llama, Mistral, Mixtral models
- Model Evaluation: Tensorboard
List of specific pre-training recipes used by the launch scripts.
Source | Model | Size | Sequence length | Nodes | Instance | Accelerator | Recipe | Script |
---|---|---|---|---|---|---|---|---|
Hugging Face | Llama 3.2 | 11b | 8192 | 4 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.2 | 90b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.2 | 1b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.2 | 3b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 70b | 16384 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 70b | 16384 | 64 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 70b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 70b | 8192 | 64 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3 | 70b | 8192 | 16 | ml.trn1.32xlarge | TRN | link | link |
Hugging Face | Llama 3.1 | 8b | 16384 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 8b | 16384 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 8b | 8192 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3.1 | 8b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Llama 3 | 8b | 8192 | 4 | ml.trn1.32xlarge | TRN | link | link |
Megatron | Llama 3.1 | 8b | 8192 | 16 | ml.p5.48xlarge | GPU H100 | link | - |
Hugging Face | Mistral | 7b | 16384 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mistral | 7b | 16384 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mistral | 7b | 8192 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mistral | 7b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 22b | 16384 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 22b | 16384 | 64 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 22b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 22b | 8192 | 64 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 7b | 16384 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 7b | 16384 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 7b | 8192 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Hugging Face | Mixtral | 7b | 8192 | 32 | ml.p5.48xlarge | GPU H100 | link | link |
List of specific fine-tuning recipes used by the launch scripts. All model sources are from Hugging Face.
Model | Method | Size | Sequence length | Nodes | Instance | Accelerator | Recipe | Script |
---|---|---|---|---|---|---|---|---|
Llama 3.1 | QLoRA | 405b | 131072 | 2 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 405b | 16384 | 6 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | QLoRA | 405b | 16384 | 2 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 405b | 8192 | 6 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | QLoRA | 405b | 8192 | 2 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | SFT | 70b | 16384 | 16 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 70b | 16384 | 2 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | SFT | 70b | 8192 | 10 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 70b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | SFT | 8b | 16384 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 8b | 16384 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | SFT | 8b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | LoRA | 8b | 8192 | 1 | ml.p5.48xlarge | GPU H100 | link | link |
Llama 3.1 | SFT | 70b | 8192 | 32 | ml.p4d.24xlarge | GPU A100 | link | link |
Llama 3.1 | LoRA | 70b | 8192 | 20 | ml.p4d.24xlarge | GPU A100 | link | link |
Llama 3.1 | SFT | 8b | 8192 | 4 | ml.p4d.24xlarge | GPU A100 | link | link |
Llama 3.1 | LoRA | 8b | 8192 | 1 | ml.p4d.24xlarge | GPU A100 | link | link |
Llama 3 | SFT | 8b | 8192 | 1 | ml.trn1.32xlarge | TRN | link | link |
Amazon SageMaker HyperPod recipes should be installed on the head node of your HyperPod cluster or on your local machine with a virtual python environment.
git clone --recursive [email protected]:aws/sagemaker-hyperpod-recipes.git
cd sagemaker-hyperpod-recipes
python3 -m venv venv
source venv/bin/activate
pip3 install -r requirements.txt
When using the SageMaker HyperPod recipes, you can either create your own training script or leverage the SageMaker HyperPod adapter, which includes popular publicly-available models like Llama or Mistral. Based on your specific needs, you might need to modify the parameters defined in the recipes for pre-training or fine-tuning. Once your configurations are setup, you can run training on SageMaker HyperPod (with Slurm or Amazon EKS) for workload orchestration. Alternatively, you can run the recipe on SageMaker training jobs using the Amazon SageMaker Python SDK.
To run a recipe via a Slurm job on a HyperPod cluster, you need to SSH into the head node of the cluster and clone the HyperPod recipes repository onto a shared filesystem, such as FSX or NFS. Next, follow the installation instructions to set up a Python virtual environment with the required dependencies. Once the environment is ready, you can launch a training job from the launcher_scripts folder. For example, you can modify the recipe launcher script run_hf_llama3_8b_seq8k_gpu_p5x16_pretrain with customized configurations such as your image or output directory. Once setting all the necessary parameters in the recipe launcher, you can start the training process by running the script.
We recommend that you utilize enroot
to initiate a training process on
the Slurm cluster. You can get the latest docker image from SMP release notes. You can refer to the following example to generate a squash file
employing the enroot
command. Please refer to the following documentation on building an AWS-optimized Nemo-Launcher image.
REGION="us-west-2"
IMAGE="658645717510.dkr.ecr.${REGION}.amazonaws.com/smdistributed-modelparallel:${TAG}"
aws ecr get-login-password --region "${REGION}" | docker login --username AWS --password-stdin 855988369404.dkr.ecr.${REGION}.amazonaws.com
enroot import -o $PWD/smdistributed-modelparallel.sqsh dockerd://${IMAGE}
mv $PWD/smdistributed-modelparallel.sqsh "/fsx/smdistributed-modelparallel.sqsh"
To use a prebuilt enroot:
wget https://sagemaker-distributed-model-parallel.s3.us-west-2.amazonaws.com/enroot/2.4.1-gpu-py311-cu121-ubuntu20.04-sagemaker-smpv2.7.0.sqsh
To use the Enroot squash file to start training, use the following example to
modify the recipes_collection/config.yaml
file.
container: /fsx/smdistributed-modelparallel.sqsh
The launcher script has variables such as TRAIN_DIR
which need to be set either by modifying the launcher script, or by setting environment variables. For example:
EXP_DIR=<your_exp_dir> TRAIN_DIR=<your_train_data_dir> VAL_DIR=<your_val_data_dir> bash ./launcher_scripts/llama/run_hf_llama3_8b_seq16k_gpu_p5x16_pretrain.sh
Prior to commencing training on your cluster, you are required to configure your local environment by adhering to the installation instructions. Additionally, you will need to install Kubectl and Helm on your local machine. Refer to the following documentation for installation of Kubectl and Helm.
You can now proceed with submitting a training job by utilizing the same launcher script with the following command:
aws eks update-kubeconfig --region "${CLUSTER_REGION}" --name "${CLUSTER_NAME}"
launcher_scripts/llama/run_hf_llama3_8b_seq8192.sh
We recommend that you utilize HyperPod command-line tool to launch a training job.
hyperpod start-job --recipe training/llama/hf_llama3_8b_seq16k_gpu_p5x16_pretrain \
--persistent-volume-claims fsx-claim:data \
--override-parameters \
'{
"recipes.run.name": "hf-llama3-8b",
"recipes.exp_manager.exp_dir": "/data/<your_exp_dir>",
"container": "658645717510.dkr.ecr.<region>.amazonaws.com/smdistributed-modelparallel:2.4.1-gpu-py311-cu121",
"recipes.model.data.train_dir": "<your_train_data_dir>",
"recipes.model.data.val_dir": "<your_val_data_dir>",
"cluster": "k8s",
"cluster_type": "k8s"
}'
SageMaker training jobs automatically spin up a resilient distributed training cluster, monitors the infrastructure, and auto-recovers from faults to ensure a smooth training experience. You can leverage the SageMaker Python SDK to execute your recipes on SageMaker training jobs.
python3 -m venv venv
source venv/bin/activate
pip3 install --upgrade pip setuptools
# install SageMaker SDK
pip install --upgrade sagemaker
The following Python code-snippet demonstrates how to submit a recipe to
run on a SageMaker training jobs by utilizing the PyTorch
estimator from the SageMaker Python SDK.
For example, to run the llama3-8b recipe on
a SageMaker training jobs, you need to set training_recipe
arg to indicate which recipe: this
can be a recipe from one of the available ones, or a url or a local yaml file containing a modified
recipe. Please also modify the local directory paths and hf access token either by providing
recipe_overrides
or by modifying the recipe yaml file directly (the url or local file).
import os
import sagemaker,boto3
from sagemaker.debugger import TensorBoardOutputConfig
from sagemaker.pytorch import PyTorch
sagemaker_session = sagemaker.Session()
role = sagemaker.get_execution_role()
bucket = sagemaker_session.default_bucket()
output = os.path.join(f"s3://{bucket}", "output")
output_path = "<s3 url>"
recipe_overrides = {
"run": {
"results_dir": "/opt/ml/model",
},
"exp_manager": {
"exp_dir": "",
"explicit_log_dir": "/opt/ml/output/tensorboard",
"checkpoint_dir": "/opt/ml/checkpoints",
},
"model": {
"data": {
"train_dir": "/opt/ml/input/data/train",
"val_dir": "/opt/ml/input/data/val",
},
},
}
tensorboard_output_config = TensorBoardOutputConfig(
s3_output_path=os.path.join(output, 'tensorboard'),
container_local_output_path=recipe_overrides["exp_manager"]["explicit_log_dir"]
)
estimator = PyTorch(
output_path=output_path,
base_job_name=f"llama-recipe",
role=role,
instance_type="ml.p5.48xlarge",
training_recipe="training/llama/hf_llama3_8b_seq8k_gpu_p5x16_pretrain",
recipe_overrides=recipe_overrides,
sagemaker_session=sagemaker_session,
tensorboard_output_config=tensorboard_output_config,
)
estimator.fit(inputs={"train": "s3 or fsx input", "val": "s3 or fsx input"}, wait=True)
Running the above code creates a PyTorch
estimator object with the specified training recipe
and then trains the model using the fit()
method. The new training_recipe
parameter enables you
to specify the recipe you want to use.
During training, if GPU memory usage approaches its limit, attempting to save sharded checkpoints to an S3 storage may result in a core dump. To address this issue, you may choose to:
- Reduce the overall memory consumption of the model training:
- Increase the number of compute nodes for the traninig process.
- Decrease the batch size
- Increase the sharding degrees, etc.
- Use FSx as the shared file system
By taking one of the above approaches, you can alleviate the memory pressure and prevent a core dump from occurring during checkpoint saving.
Llama 3.2 specifically requires transformers version 4.45.2 or above. They should be installed automatically in the container during job launch if using slurm or k8s. If not, you can update your requirements.txt or container so that transformers==4.45.2 is installed.
Follow the instructions on the "Installing" then use the following command to install the dependencies for testing:
pip install pytest
pip install pytest-cov
To run the unit tests, navigate to the root directory and use the command
python -m pytest
plus any desired flags.
The pyproject.toml
file defines additional options that are always appended to the pytest
command:
[tool.pytest.ini_options]
...
addopts = [
"--cache-clear",
"--quiet",
"--durations=0",
"--cov=launcher/",
# uncomment this line to see a detailed HTML test coverage report instead of the usual summary table output to stdout.
# "--cov-report=html",
"tests/",
]
We use pre-commit to unify our coding format, steps to setup as as follows:
- Install pre-commit which helps us run formatters before commit using
pip install pre-commit
- Setup hooks from our pre-commit hook configs in
.pre-commit-config.yaml
usingpre-commit install
When you commit, pre-commit hooks will be applied. If for some reason you need to skip the check, you can rungit commit ... --no-verify
but make sure to include the reason to skip pre-commit in the commit message.
See CONTRIBUTING for more information.
This project is licensed under the Apache-2.0 License.