@@ -112,7 +112,6 @@ def test_base_model_evaluation_uses_correct_weights(self):
112112 builtin_metrics = TEST_CONFIG ["builtin_metrics" ],
113113 custom_metrics = TEST_CONFIG ["custom_metrics_json" ],
114114 s3_output_path = TEST_CONFIG ["s3_output_path" ],
115- mlflow_resource_arn = TEST_CONFIG ["mlflow_tracking_server_arn" ],
116115 evaluate_base_model = TEST_CONFIG ["evaluate_base_model" ],
117116 )
118117
@@ -216,14 +215,19 @@ def test_base_model_evaluation_uses_correct_weights(self):
216215 if execution .status .failure_reason :
217216 logger .error (f" Failure reason: { execution .status .failure_reason } " )
218217
219- # Log step failures
218+ # Log step failures with detailed information
220219 if execution .status .step_details :
221- logger .error ("\n Failed steps:" )
220+ logger .error ("\n " + "=" * 80 )
221+ logger .error ("DETAILED STEP FAILURE INFORMATION:" )
222+ logger .error ("=" * 80 )
222223 for step in execution .status .step_details :
223- if "failed" in step .status .lower ():
224- logger .error (f" { step .name } : { step .status } " )
225- if step .failure_reason :
226- logger .error (f" Reason: { step .failure_reason } " )
224+ logger .error (f"\n Step: { step .name } " )
225+ logger .error (f" Status: { step .status } " )
226+ logger .error (f" Start Time: { step .start_time } " )
227+ logger .error (f" End Time: { step .end_time } " )
228+ if step .failure_reason :
229+ logger .error (f" ❌ FAILURE REASON: { step .failure_reason } " )
230+ logger .error ("=" * 80 )
227231
228232 # Re-raise to fail the test
229233 raise
@@ -248,7 +252,6 @@ def test_base_model_false_still_works(self):
248252 dataset = TEST_CONFIG ["dataset_s3_uri" ],
249253 builtin_metrics = TEST_CONFIG ["builtin_metrics" ],
250254 s3_output_path = TEST_CONFIG ["s3_output_path" ],
251- mlflow_resource_arn = TEST_CONFIG ["mlflow_tracking_server_arn" ],
252255 evaluate_base_model = False , # Only evaluate custom model
253256 )
254257
0 commit comments