Skip to content

Commit 53448a4

Browse files
committed
Skip unsupported test, xfail capcity and resource error
1 parent 4fd497c commit 53448a4

File tree

4 files changed

+29
-4
lines changed

4 files changed

+29
-4
lines changed

tests/integ/sagemaker/jumpstart/constants.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,6 @@ def _to_s3_path(filename: str, s3_prefix: Optional[str]) -> str:
5353
("meta-textgeneration-llama-2-7b", "2.*"): ("training-datasets/sec_amazon/"),
5454
("meta-textgeneration-llama-2-7b", "3.*"): ("training-datasets/sec_amazon/"),
5555
("meta-textgeneration-llama-2-7b", "4.*"): ("training-datasets/sec_amazon/"),
56-
("huggingface-llm-gemma-7b", "*"): ("training-datasets/genuq/small/"),
5756
("meta-textgenerationneuron-llama-2-7b", "*"): ("training-datasets/sec_amazon/"),
5857
}
5958

tests/integ/sagemaker/jumpstart/model/test_jumpstart_model.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@
3434
download_inference_assets,
3535
get_sm_session,
3636
get_tabular_data,
37+
x_fail_if_ice,
3738
)
3839

3940
INF2_SUPPORTED_REGIONS = {
@@ -191,7 +192,10 @@ def test_jumpstart_gated_model(setup):
191192
assert response is not None
192193

193194

195+
@x_fail_if_ice
194196
def test_jumpstart_gated_model_inference_component_enabled(setup):
197+
# x_fail_if_ice marks this test as xfail on CapacityError — ml.g5.2xlarge capacity
198+
# is shared across parallel CI runs and may be transiently exhausted.
195199

196200
model_id = "meta-textgeneration-llama-2-7b"
197201

tests/integ/sagemaker/jumpstart/private_hub/estimator/test_jumpstart_private_hub_estimator.py

Lines changed: 20 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@
3838

3939
TEST_MODEL_IDS = {
4040
"huggingface-spc-bert-base-cased",
41-
"huggingface-llm-gemma-7b",
4241
"catboost-regression-model",
4342
}
4443

@@ -136,9 +135,18 @@ def test_jumpstart_hub_estimator_with_session(setup, add_model_references):
136135
assert response is not None
137136

138137

138+
@pytest.mark.skip(
139+
reason=(
140+
"meta-textgeneration-llama-2-7b has been removed from the SageMaker public JumpStart hub. "
141+
"Gated model EULA enforcement is covered by test_jumpstart_hub_gated_model in "
142+
"test_jumpstart_private_hub_model.py and test_gated_model_training_v1/v2 in "
143+
"test_jumpstart_estimator.py. TODO: replace with a suitable gated model that supports "
144+
"training via private hub without requiring specific VPC endpoint configuration."
145+
)
146+
)
139147
def test_jumpstart_hub_gated_estimator_with_eula(setup, add_model_references):
140148

141-
model_id, model_version = "huggingface-llm-gemma-7b", "*"
149+
model_id, model_version = "meta-textgeneration-llama-2-7b", "*"
142150

143151
estimator = JumpStartEstimator(
144152
model_id=model_id,
@@ -170,9 +178,18 @@ def test_jumpstart_hub_gated_estimator_with_eula(setup, add_model_references):
170178
assert response is not None
171179

172180

181+
@pytest.mark.skip(
182+
reason=(
183+
"meta-textgeneration-llama-2-7b has been removed from the SageMaker public JumpStart hub. "
184+
"Gated model EULA enforcement is covered by test_jumpstart_hub_gated_model in "
185+
"test_jumpstart_private_hub_model.py and test_gated_model_training_v1/v2 in "
186+
"test_jumpstart_estimator.py. TODO: replace with a suitable gated model that supports "
187+
"training via private hub without requiring specific VPC endpoint configuration."
188+
)
189+
)
173190
def test_jumpstart_hub_gated_estimator_without_eula(setup, add_model_references):
174191

175-
model_id, model_version = "huggingface-llm-gemma-7b", "*"
192+
model_id, model_version = "meta-textgeneration-llama-2-7b", "*"
176193

177194
estimator = JumpStartEstimator(
178195
model_id=model_id,

tests/integ/sagemaker/serve/test_serve_model_builder_inference_component_happy.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -105,6 +105,11 @@ def test_model_builder_ic_sagemaker_endpoint(
105105
if caught_ex:
106106
logger.exception(caught_ex)
107107
cleanup_resources(sagemaker_session, [LLAMA_IC_NAME])
108+
if "ResourceLimitExceeded" in str(caught_ex) or "CapacityError" in str(caught_ex):
109+
# Mark as xfail rather than hard-failing — ml.g5.24xlarge capacity is shared
110+
# across parallel CI runs and may be transiently exhausted. This is not a
111+
# code regression; the test should be retried when capacity is available.
112+
pytest.xfail(str(caught_ex))
108113
assert False, f"{caught_ex} thrown when running mb-IC deployment test."
109114

110115
cleanup_resources(sagemaker_session, [LLAMA_IC_NAME])

0 commit comments

Comments
 (0)