Skip to content

infra: rename canary_quick pytest mark to release #2167

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Feb 19, 2021
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion tests/integ/test_artifact_analytics.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ def generate_artifacts(sagemaker_session):
_delete_resources(artifacts)


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip("Failing as restricted to the SageMaker/Pipeline runtimes")
def test_artifact_analytics(sagemaker_session):
with generate_artifacts(sagemaker_session):
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_auto_ml.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@
tests.integ.test_region() in tests.integ.NO_AUTO_ML_REGIONS,
reason="AutoML is not supported in the region yet.",
)
@pytest.mark.canary_quick
@pytest.mark.release
def test_auto_ml_fit(sagemaker_session):
auto_ml = AutoML(
role=ROLE,
Expand Down Expand Up @@ -260,7 +260,7 @@ def test_best_candidate(sagemaker_session):
tests.integ.test_region() in tests.integ.NO_AUTO_ML_REGIONS,
reason="AutoML is not supported in the region yet.",
)
@pytest.mark.canary_quick
@pytest.mark.release
def test_deploy_best_candidate(sagemaker_session, cpu_instance_type):
auto_ml_utils.create_auto_ml_job_if_not_exist(sagemaker_session)

Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_byo_estimator.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ def serialize(self, data):
return json.dumps(js)


@pytest.mark.canary_quick
@pytest.mark.release
def test_byo_estimator(sagemaker_session, region, cpu_instance_type, training_set):
"""Use Factorization Machines algorithm as an example here.

Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_experiments_analytics.py
Original file line number Diff line number Diff line change
Expand Up @@ -108,7 +108,7 @@ def experiment_with_artifacts(sagemaker_session):
_delete_resources(sm, experiment_name, trials)


@pytest.mark.canary_quick
@pytest.mark.release
def test_experiment_analytics_artifacts(sagemaker_session):
with experiment_with_artifacts(sagemaker_session) as experiment_name:
analytics = ExperimentAnalytics(
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_horovod.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
horovod_dir = os.path.join(os.path.dirname(__file__), "..", "data", "horovod")


@pytest.mark.canary_quick
@pytest.mark.release
def test_hvd_cpu(
sagemaker_session,
tensorflow_training_latest_version,
Expand All @@ -45,7 +45,7 @@ def test_hvd_cpu(
)


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skipif(
integ.test_region() in integ.TRAINING_NO_P2_REGIONS, reason="no ml.p2 instances in this region"
)
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_horovod_mx.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
horovod_dir = os.path.join(os.path.dirname(__file__), "..", "data", "horovod")


@pytest.mark.canary_quick
@pytest.mark.release
def test_hvd_cpu(
mxnet_training_latest_version,
mxnet_training_latest_py_version,
Expand All @@ -45,7 +45,7 @@ def test_hvd_cpu(
)


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skipif(
integ.test_region() in integ.TRAINING_NO_P2_REGIONS, reason="no ml.p2 instances in this region"
)
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_inference_pipeline.py
Original file line number Diff line number Diff line change
Expand Up @@ -90,7 +90,7 @@ def test_inference_pipeline_batch_transform(sagemaker_session, cpu_instance_type
transformer.wait()


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip(
reason="This test has always failed, but the failure was masked by a bug. "
"This test should be fixed. Details in https://github.com/aws/sagemaker-python-sdk/pull/968"
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_linear_learner.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ def training_set():
return datasets.one_p_mnist()


@pytest.mark.canary_quick
@pytest.mark.release
def test_linear_learner(sagemaker_session, cpu_instance_type, training_set):
job_name = unique_name_from_base("linear-learner")

Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_marketplace.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@
)


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skipif(
tests.integ.test_region() in tests.integ.NO_MARKET_PLACE_REGIONS,
reason="Marketplace is not available in {}".format(tests.integ.test_region()),
Expand Down Expand Up @@ -150,7 +150,7 @@ def test_marketplace_attach(sagemaker_session, cpu_instance_type):
print(predictor.predict(test_x.values).decode("utf-8"))


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skipif(
tests.integ.test_region() in tests.integ.NO_MARKET_PLACE_REGIONS,
reason="Marketplace is not available in {}".format(tests.integ.test_region()),
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_model_monitor.py
Original file line number Diff line number Diff line change
Expand Up @@ -271,7 +271,7 @@ def updated_output_kms_key(sagemaker_session):
tests.integ.test_region() in tests.integ.NO_MODEL_MONITORING_REGIONS,
reason="ModelMonitoring is not yet supported in this region.",
)
@pytest.mark.canary_quick
@pytest.mark.release
def test_default_monitor_suggest_baseline_and_create_monitoring_schedule_with_customizations(
sagemaker_session, output_kms_key, volume_kms_key, predictor
):
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_mxnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ def mxnet_training_job(
return mx.latest_training_job.name


@pytest.mark.canary_quick
@pytest.mark.release
def test_attach_deploy(mxnet_training_job, sagemaker_session, cpu_instance_type):
endpoint_name = "test-mxnet-attach-deploy-{}".format(sagemaker_timestamp())

Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_neo_mxnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ def mxnet_training_job(
return mx.latest_training_job.name


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip(
reason="This test is failing because the image uri and the training script format has changed."
)
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_ntm.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
from tests.integ.record_set import prepare_record_set_from_local_files


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip(
reason="This test has always failed, but the failure was masked by a bug. "
"This test should be fixed. Details in https://github.com/aws/sagemaker-python-sdk/pull/968"
Expand Down
6 changes: 3 additions & 3 deletions tests/integ/test_processing.py
Original file line number Diff line number Diff line change
Expand Up @@ -153,7 +153,7 @@ def test_sklearn(sagemaker_session, sklearn_latest_version, cpu_instance_type):
assert ROLE in job_description["RoleArn"]


@pytest.mark.canary_quick
@pytest.mark.release
def test_sklearn_with_customizations(
sagemaker_session, image_uri, sklearn_latest_version, cpu_instance_type, output_kms_key
):
Expand Down Expand Up @@ -369,7 +369,7 @@ def test_sklearn_with_no_inputs_or_outputs(
assert job_description["StoppingCondition"] == {"MaxRuntimeInSeconds": 3600}


@pytest.mark.canary_quick
@pytest.mark.release
def test_script_processor(sagemaker_session, image_uri, cpu_instance_type, output_kms_key):
input_file_path = os.path.join(DATA_DIR, "dummy_input.txt")

Expand Down Expand Up @@ -528,7 +528,7 @@ def test_script_processor_with_no_inputs_or_outputs(
assert job_description["StoppingCondition"] == {"MaxRuntimeInSeconds": 3600}


@pytest.mark.canary_quick
@pytest.mark.release
def test_processor(sagemaker_session, image_uri, cpu_instance_type, output_kms_key):
script_path = os.path.join(DATA_DIR, "dummy_script.py")

Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_pytorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,7 @@ def fixture_training_job_with_latest_inference_version(
return pytorch.latest_training_job.name


@pytest.mark.canary_quick
@pytest.mark.release
def test_fit_deploy(
pytorch_training_job_with_latest_infernce_version, sagemaker_session, cpu_instance_type
):
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_rl.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
from tests.integ.timeout import timeout, timeout_and_delete_endpoint_by_name


@pytest.mark.canary_quick
@pytest.mark.release
def test_coach_mxnet(sagemaker_session, coach_mxnet_latest_version, cpu_instance_type):
estimator = _test_coach(
sagemaker_session, RLFramework.MXNET, coach_mxnet_latest_version, cpu_instance_type
Expand Down Expand Up @@ -106,7 +106,7 @@ def _test_coach(sagemaker_session, rl_framework, rl_coach_version, cpu_instance_
test_region() not in RL_SUPPORTED_REGIONS,
reason="Updated RL images aren't in {}".format(test_region()),
)
@pytest.mark.canary_quick
@pytest.mark.release
def test_ray_tf(sagemaker_session, ray_tensorflow_latest_version, cpu_instance_type):
source_dir = os.path.join(DATA_DIR, "ray_cartpole")
cartpole = "train_ray.py"
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_sklearn.py
Original file line number Diff line number Diff line change
Expand Up @@ -111,7 +111,7 @@ def test_training_with_network_isolation(
]


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip(
reason="This test has always failed, but the failure was masked by a bug. "
"This test should be fixed. Details in https://github.com/aws/sagemaker-python-sdk/pull/968"
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_sparkml_serving.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
from tests.integ.timeout import timeout_and_delete_endpoint_by_name


@pytest.mark.canary_quick
@pytest.mark.release
def test_sparkml_model_deploy(sagemaker_session, cpu_instance_type):
# Uploads an MLeap serialized MLeap model to S3 and use that to deploy
# a SparkML model to perform inference
Expand Down
2 changes: 1 addition & 1 deletion tests/integ/test_tf.py
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ def test_server_side_encryption(sagemaker_session, tf_full_version, tf_full_py_v
)


@pytest.mark.canary_quick
@pytest.mark.release
def test_mnist_distributed(
sagemaker_session,
instance_type,
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_tfs.py
Original file line number Diff line number Diff line change
Expand Up @@ -138,7 +138,7 @@ def tfs_predictor_with_accelerator(
yield predictor


@pytest.mark.canary_quick
@pytest.mark.release
def test_predict(tfs_predictor):
input_data = {"instances": [1.0, 2.0, 5.0]}
expected_result = {"predictions": [3.5, 4.0, 5.5]}
Expand All @@ -151,7 +151,7 @@ def test_predict(tfs_predictor):
tests.integ.test_region() not in tests.integ.EI_SUPPORTED_REGIONS,
reason="EI is not supported in region {}".format(tests.integ.test_region()),
)
@pytest.mark.canary_quick
@pytest.mark.release
def test_predict_with_accelerator(tfs_predictor_with_accelerator):
input_data = {"instances": [1.0, 2.0, 5.0]}
expected_result = {"predictions": [3.5, 4.0, 5.5]}
Expand Down
4 changes: 2 additions & 2 deletions tests/integ/test_transformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ def mxnet_transform_input(sagemaker_session):
)


@pytest.mark.canary_quick
@pytest.mark.release
def test_transform_mxnet(
mxnet_estimator, mxnet_transform_input, sagemaker_session, cpu_instance_type
):
Expand Down Expand Up @@ -108,7 +108,7 @@ def test_transform_mxnet(
assert input_filter == job_desc["DataProcessing"]["InputFilter"]


@pytest.mark.canary_quick
@pytest.mark.release
def test_attach_transform_kmeans(sagemaker_session, cpu_instance_type):
kmeans = KMeans(
role="SageMakerRole",
Expand Down
12 changes: 6 additions & 6 deletions tests/integ/test_tuner.py
Original file line number Diff line number Diff line change
Expand Up @@ -158,7 +158,7 @@ def _tune(
return tuner


@pytest.mark.canary_quick
@pytest.mark.release
def test_tuning_kmeans(
sagemaker_session, kmeans_train_set, kmeans_estimator, hyperparameter_ranges, cpu_instance_type
):
Expand Down Expand Up @@ -525,7 +525,7 @@ def test_stop_tuning_job(sagemaker_session, cpu_instance_type):


@pytest.mark.slow_test
@pytest.mark.canary_quick
@pytest.mark.release
def test_tuning_mxnet(
sagemaker_session,
mxnet_training_latest_version,
Expand Down Expand Up @@ -578,7 +578,7 @@ def test_tuning_mxnet(
predictor.predict(data)


@pytest.mark.canary_quick
@pytest.mark.release
def test_tuning_tf(
sagemaker_session,
cpu_instance_type,
Expand Down Expand Up @@ -675,7 +675,7 @@ def test_tuning_tf_vpc_multi(
tuner.fit(inputs, job_name=tuning_job_name)


@pytest.mark.canary_quick
@pytest.mark.release
def test_tuning_chainer(
sagemaker_session, chainer_latest_version, chainer_latest_py_version, cpu_instance_type
):
Expand Down Expand Up @@ -742,7 +742,7 @@ def test_tuning_chainer(
assert len(output) == batch_size


@pytest.mark.canary_quick
@pytest.mark.release
@pytest.mark.skip(
reason="This test has always failed, but the failure was masked by a bug. "
"This test should be fixed. Details in https://github.com/aws/sagemaker-python-sdk/pull/968"
Expand Down Expand Up @@ -817,7 +817,7 @@ def test_attach_tuning_pytorch(
_assert_model_name_match(sagemaker_session.sagemaker_client, endpoint_name, model_name)


@pytest.mark.canary_quick
@pytest.mark.release
def test_tuning_byo_estimator(sagemaker_session, cpu_instance_type):
"""Use Factorization Machines algorithm as an example here.

Expand Down