Skip to content

Commit a1b10a9

Browse files
committed
Add cleanup mechanism for MC dataset integ test
1 parent 7732ecf commit a1b10a9

3 files changed

Lines changed: 9 additions & 12 deletions

File tree

sagemaker-train/src/sagemaker/ai_registry/dataset.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -389,7 +389,6 @@ def get_versions(self) -> List["DataSet"]:
389389

390390
return datasets
391391

392-
@classmethod
393392
@classmethod
394393
@_telemetry_emitter(feature=Feature.MODEL_CUSTOMIZATION, func_name="DataSet.get_all")
395394
def get_all(cls, max_results: Optional[int] = None, sagemaker_session=None):

sagemaker-train/tests/integ/ai_registry/conftest.py

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -106,13 +106,8 @@ def cleanup_list():
106106
"""Track resources for cleanup."""
107107
resources = []
108108
yield resources
109-
for evaluator in resources:
109+
for resource in resources:
110110
try:
111-
from sagemaker.ai_registry.air_hub import AIRHub
112-
AIRHub.delete_hub_content(
113-
hub_content_type=evaluator.hub_content_type,
114-
hub_content_name=evaluator.name,
115-
hub_content_version=evaluator.version
116-
)
111+
resource.delete()
117112
except Exception:
118113
pass

sagemaker-train/tests/integ/ai_registry/test_dataset.py

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -129,9 +129,10 @@ def test_create_dataset_from_s3_nova_eval(self, unique_name, test_bucket, cleanu
129129
cleanup_list.append(dataset)
130130
assert dataset.name == unique_name
131131

132-
def test_get_dataset(self, unique_name, sample_jsonl_file):
132+
def test_get_dataset(self, unique_name, sample_jsonl_file, cleanup_list):
133133
"""Test retrieving dataset by name."""
134134
created = DataSet.create(name=unique_name, source=sample_jsonl_file, wait=False)
135+
cleanup_list.append(created)
135136
retrieved = DataSet.get(unique_name)
136137
assert retrieved.name == created.name
137138
assert retrieved.arn == created.arn
@@ -141,16 +142,18 @@ def test_get_all_datasets(self):
141142
datasets = list(DataSet.get_all(max_results=5))
142143
assert isinstance(datasets, list)
143144

144-
def test_dataset_refresh(self, unique_name, sample_jsonl_file):
145+
def test_dataset_refresh(self, unique_name, sample_jsonl_file, cleanup_list):
145146
"""Test refreshing dataset status."""
146147
dataset = DataSet.create(name=unique_name, source=sample_jsonl_file, wait=False)
148+
cleanup_list.append(dataset)
147149
dataset.refresh()
148150
time.sleep(3)
149151
assert dataset.status in [HubContentStatus.IMPORTING.value, HubContentStatus.AVAILABLE.value]
150152

151-
def test_dataset_get_versions(self, unique_name, sample_jsonl_file):
153+
def test_dataset_get_versions(self, unique_name, sample_jsonl_file, cleanup_list):
152154
"""Test getting dataset versions."""
153155
dataset = DataSet.create(name=unique_name, source=sample_jsonl_file, wait=False)
156+
cleanup_list.append(dataset)
154157
versions = dataset.get_versions()
155158
assert len(versions) >= 1
156159
assert all(isinstance(v, DataSet) for v in versions)
@@ -178,7 +181,7 @@ def test_create_dataset_version(self, unique_name, sample_jsonl_file, cleanup_li
178181
"""Test creating new dataset version."""
179182
dataset = DataSet.create(name=unique_name, source=sample_jsonl_file, wait=False)
180183
result = dataset.create_version(sample_jsonl_file)
181-
cleanup_list.append(cleanup_list)
184+
cleanup_list.append(dataset)
182185
assert result is True
183186

184187
def test_dataset_validation_invalid_extension(self, unique_name):

0 commit comments

Comments
 (0)