Skip to content

Commit 81ba403

Browse files
authored
Merge branch 'master' into ptddp-launcher
2 parents a4352c6 + b4f05b8 commit 81ba403

39 files changed

+3617
-415
lines changed

CHANGELOG.md

+23
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,28 @@
11
# Changelog
22

3+
## v2.100.0 (2022-07-18)
4+
5+
### Features
6+
7+
* upgrade to support python 3.10
8+
* Add target_model to support multi-model endpoints
9+
* Added support for feature group schema change and feature parameters
10+
11+
### Bug Fixes and Other Changes
12+
13+
* enable model.register without 'inference' & 'transform' instances
14+
* rename RegisterModel inner steps to prevent duplicate step names
15+
* remove primitive_or_expr() from conditions
16+
* support pipeline variables for spark processors run arguments
17+
* make 'ModelInput' field optional for inference recommendation
18+
* Fix processing image uri param
19+
* fix: neo inferentia as compilation target not using framework ver
20+
21+
### Documentation Changes
22+
23+
* SageMaker model parallel library v1.10.0 documentation
24+
* add detail & links to clarify docstrings
25+
326
## v2.99.0 (2022-07-08)
427

528
### Features

VERSION

+1-1
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.99.1.dev0
1+
2.100.1.dev0

doc/api/training/smd_model_parallel_general.rst

+10
Original file line numberDiff line numberDiff line change
@@ -178,6 +178,16 @@ PyTorch-specific Parameters
178178
- 1
179179
- The number of devices over which the tensor parallel modules will be distributed.
180180
If ``tensor_parallel_degree`` is greater than 1, then ``ddp`` must be set to ``True``.
181+
* - ``fp16`` (**smdistributed-modelparallel**>=v1.10)
182+
- bool
183+
- ``False``
184+
- To run FP16 training, add ``"fp16"'": True`` to the smp configuration.
185+
Other APIs remain the same between FP16 and FP32.
186+
If ``fp16`` is enabled and when user calls ``smp.DistributedModel``,
187+
the model will be wrapped with ``FP16_Module``, which converts the model
188+
to FP16 dtype and deals with forward pass in FP16.
189+
If ``fp16`` is enabled and when user calls ``smp.DistributedOptimizer``,
190+
the optimizer will be wrapped with ``FP16_Optimizer``.
181191
* - ``fp16_params`` (**smdistributed-modelparallel**>=v1.6)
182192
- bool
183193
- ``False``

doc/api/training/smp_versions/archives.rst

+1
Original file line numberDiff line numberDiff line change
@@ -3,6 +3,7 @@
33
.. toctree::
44
:maxdepth: 1
55

6+
v1_9_0.rst
67
v1_6_0.rst
78
v1_5_0.rst
89
v1_4_0.rst

doc/api/training/smp_versions/latest.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ depending on which version of the library you need to use.
1010
To use the library, reference the
1111
**Common API** documentation alongside the framework specific API documentation.
1212

13-
Version 1.7.0, 1.8.0, 1.8.1, 1.9.0 (Latest)
13+
Version 1.10.0 (Latest)
1414
===========================================
1515

1616
To use the library, reference the Common API documentation alongside the framework specific API documentation.

doc/api/training/smp_versions/latest/smd_model_parallel_pytorch.rst

+267-67
Large diffs are not rendered by default.

doc/api/training/smp_versions/latest/smd_model_parallel_pytorch_tensor_parallel.rst

+111-85
Large diffs are not rendered by default.

0 commit comments

Comments
 (0)