diff --git a/doc/api/training/distributed.rst b/doc/api/training/distributed.rst index 44ea860861..21837bc1e4 100644 --- a/doc/api/training/distributed.rst +++ b/doc/api/training/distributed.rst @@ -10,7 +10,7 @@ The SageMaker Distributed Data Parallel Library ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ .. toctree:: - :maxdepth: 3 + :maxdepth: 2 smd_data_parallel sdp_versions/latest @@ -23,7 +23,7 @@ The SageMaker Distributed Model Parallel Library ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ .. toctree:: - :maxdepth: 3 + :maxdepth: 2 smd_model_parallel smp_versions/latest diff --git a/doc/api/training/smd_model_parallel_release_notes/smd_model_parallel_change_log.rst b/doc/api/training/smd_model_parallel_release_notes/smd_model_parallel_change_log.rst index 65c63b3a83..2e5ad2a8ac 100644 --- a/doc/api/training/smd_model_parallel_release_notes/smd_model_parallel_change_log.rst +++ b/doc/api/training/smd_model_parallel_release_notes/smd_model_parallel_change_log.rst @@ -5,9 +5,48 @@ Release Notes New features, bug fixes, and improvements are regularly made to the SageMaker distributed model parallel library. -SageMaker Distributed Model Parallel 1.7.0 Release Notes +SageMaker Distributed Model Parallel 1.8.0 Release Notes ======================================================== +*Date: March. 23. 2022* + +**New Features** + +* Added tensor parallelism support for the `GPT-J model + `_. + When using the GPT-J model of Hugging Face Transformers v4.17.0 with + tensor parallelism, the SageMaker model parallel library automatically + replaces the model with a tensor parallel distributed GPT-J model. + For more information, see `Support for Hugging Face Transformer Models + `_ + in the *Amazon SageMaker Model Parallel Training developer guide*. + +**Migration to AWS Deep Learning Containers** + +This version passed benchmark testing and is migrated to the following AWS Deep Learning Containers: + +* HuggingFace 4.17.0 DLC with PyTorch 1.10.2 + + .. code:: + + 763104351884.dkr.ecr.us-west-2.amazonaws.com/huggingface-pytorch-training:1.10.2-transformers4.17.0-gpu-py38-cu113-ubuntu20.04 + + + The binary file of this version of the library for custom container users: + + .. code:: + + https://sagemaker-distributed-model-parallel.s3.us-west-2.amazonaws.com/pytorch-1.10.0/build-artifacts/2022-03-12-00-33/smdistributed_modelparallel-1.8.0-cp38-cp38-linux_x86_64.whl + + +---- + +Release History +=============== + +SageMaker Distributed Model Parallel 1.7.0 Release Notes +-------------------------------------------------------- + *Date: March. 07. 2022* **Currency Updates** @@ -49,11 +88,6 @@ This version passed benchmark testing and is migrated to the following AWS Deep 763104351884.dkr.ecr..amazonaws.com/pytorch-training:1.10.2-gpu-py38-cu113-ubuntu20.04-sagemaker ----- - -Release History -=============== - SageMaker Distributed Model Parallel 1.6.0 Release Notes -------------------------------------------------------- diff --git a/doc/api/training/smp_versions/latest.rst b/doc/api/training/smp_versions/latest.rst index da44e4f1dd..425825054b 100644 --- a/doc/api/training/smp_versions/latest.rst +++ b/doc/api/training/smp_versions/latest.rst @@ -10,8 +10,8 @@ depending on which version of the library you need to use. To use the library, reference the **Common API** documentation alongside the framework specific API documentation. -Version 1.7.0 (Latest) -====================== +Version 1.7.0, 1.8.0 (Latest) +============================= To use the library, reference the Common API documentation alongside the framework specific API documentation.