Skip to content

Commit b40d702

Browse files
committed
added skip tests
1 parent f737cb1 commit b40d702

5 files changed

Lines changed: 44 additions & 3 deletions

File tree

tests/test_flows/test_flow.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -180,6 +180,7 @@ def test_to_xml_from_xml(self):
180180
openml.flows.functions.assert_flows_equal(new_flow, flow)
181181
assert new_flow is not flow
182182

183+
@pytest.mark.skip(reason="Pending resolution of #1657")
183184
@pytest.mark.sklearn()
184185
@pytest.mark.test_server()
185186
def test_publish_flow(self):
@@ -222,6 +223,7 @@ def test_publish_existing_flow(self, flow_exists_mock):
222223
f"collected from {__file__.split('/')[-1]}: {flow.flow_id}",
223224
)
224225

226+
@pytest.mark.skip(reason="Pending resolution of #1657")
225227
@pytest.mark.sklearn()
226228
@pytest.mark.test_server()
227229
def test_publish_flow_with_similar_components(self):
@@ -273,6 +275,7 @@ def test_publish_flow_with_similar_components(self):
273275
TestBase._mark_entity_for_removal("flow", flow3.flow_id, flow3.name)
274276
TestBase.logger.info(f"collected from {__file__.split('/')[-1]}: {flow3.flow_id}")
275277

278+
@pytest.mark.skip(reason="Pending resolution of #1657")
276279
@pytest.mark.sklearn()
277280
@pytest.mark.test_server()
278281
def test_semi_legal_flow(self):
@@ -383,6 +386,7 @@ def get_sentinel():
383386
flow_id = openml.flows.flow_exists(name, version)
384387
assert not flow_id
385388

389+
@pytest.mark.skip(reason="Pending resolution of #1657")
386390
@pytest.mark.sklearn()
387391
@pytest.mark.test_server()
388392
def test_existing_flow_exists(self):
@@ -424,6 +428,7 @@ def test_existing_flow_exists(self):
424428
)
425429
assert downloaded_flow_id == flow.flow_id
426430

431+
@pytest.mark.skip(reason="Pending resolution of #1657")
427432
@pytest.mark.sklearn()
428433
@pytest.mark.test_server()
429434
def test_sklearn_to_upload_to_flow(self):

tests/test_flows/test_flow_functions.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -309,6 +309,7 @@ def test_get_flow1(self):
309309
flow = openml.flows.get_flow(1)
310310
assert flow.external_version is None
311311

312+
@pytest.mark.skip(reason="Pending resolution of #1657")
312313
@pytest.mark.sklearn()
313314
@pytest.mark.test_server()
314315
def test_get_flow_reinstantiate_model(self):
@@ -392,6 +393,7 @@ def test_get_flow_reinstantiate_flow_not_strict_pre_023(self):
392393
assert flow.flow_id is None
393394
assert "sklearn==0.19.1" not in flow.dependencies
394395

396+
@pytest.mark.skip(reason="Pending resolution of #1657")
395397
@pytest.mark.sklearn()
396398
@pytest.mark.test_server()
397399
def test_get_flow_id(self):

tests/test_runs/test_run.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -118,6 +118,7 @@ def _check_array(array, type_):
118118
else:
119119
assert run_prime_trace_content is None
120120

121+
@pytest.mark.skip(reason="Pending resolution of #1657")
121122
@pytest.mark.sklearn()
122123
@pytest.mark.test_server()
123124
def test_to_from_filesystem_vanilla(self):
@@ -153,6 +154,7 @@ def test_to_from_filesystem_vanilla(self):
153154
f"collected from {__file__.split('/')[-1]}: {run_prime.run_id}",
154155
)
155156

157+
@pytest.mark.skip(reason="Pending resolution of #1657")
156158
@pytest.mark.sklearn()
157159
@pytest.mark.flaky()
158160
@pytest.mark.test_server()
@@ -189,6 +191,7 @@ def test_to_from_filesystem_search(self):
189191
f"collected from {__file__.split('/')[-1]}: {run_prime.run_id}",
190192
)
191193

194+
@pytest.mark.skip(reason="Pending resolution of #1657")
192195
@pytest.mark.sklearn()
193196
@pytest.mark.test_server()
194197
def test_to_from_filesystem_no_model(self):
@@ -295,6 +298,7 @@ def assert_run_prediction_data(task, run, model):
295298
assert_method(y_pred, saved_y_pred)
296299
assert_method(y_test, saved_y_test)
297300

301+
@pytest.mark.skip(reason="Pending resolution of #1657")
298302
@pytest.mark.sklearn()
299303
@pytest.mark.test_server()
300304
def test_publish_with_local_loaded_flow(self):
@@ -339,6 +343,7 @@ def test_publish_with_local_loaded_flow(self):
339343
assert openml.flows.flow_exists(flow.name, flow.external_version)
340344
openml.runs.get_run(loaded_run.run_id)
341345

346+
@pytest.mark.skip(reason="Pending resolution of #1657")
342347
@pytest.mark.sklearn()
343348
@pytest.mark.test_server()
344349
def test_offline_and_online_run_identical(self):

tests/test_runs/test_run_functions.py

Lines changed: 28 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -397,6 +397,7 @@ def _check_sample_evaluations(
397397
assert evaluation > 0
398398
assert evaluation < max_time_allowed
399399

400+
@pytest.mark.skip(reason="Pending resolution of #1657")
400401
@pytest.mark.sklearn()
401402
@pytest.mark.test_server()
402403
def test_run_regression_on_classif_task(self):
@@ -414,6 +415,7 @@ def test_run_regression_on_classif_task(self):
414415
task=task,
415416
)
416417

418+
@pytest.mark.skip(reason="Pending resolution of #1657")
417419
@pytest.mark.sklearn()
418420
@pytest.mark.test_server()
419421
def test_check_erronous_sklearn_flow_fails(self):
@@ -627,6 +629,7 @@ def _run_and_upload_regression(
627629
sentinel=sentinel,
628630
)
629631

632+
@pytest.mark.skip(reason="Pending resolution of #1657")
630633
@pytest.mark.sklearn()
631634
@pytest.mark.test_server()
632635
def test_run_and_upload_logistic_regression(self):
@@ -636,6 +639,7 @@ def test_run_and_upload_logistic_regression(self):
636639
n_test_obs = self.TEST_SERVER_TASK_SIMPLE["n_test_obs"]
637640
self._run_and_upload_classification(lr, task_id, n_missing_vals, n_test_obs, "62501")
638641

642+
@pytest.mark.skip(reason="Pending resolution of #1657")
639643
@pytest.mark.sklearn()
640644
@pytest.mark.test_server()
641645
def test_run_and_upload_linear_regression(self):
@@ -667,6 +671,7 @@ def test_run_and_upload_linear_regression(self):
667671
n_test_obs = self.TEST_SERVER_TASK_REGRESSION["n_test_obs"]
668672
self._run_and_upload_regression(lr, task_id, n_missing_vals, n_test_obs, "62501")
669673

674+
@pytest.mark.skip(reason="Pending resolution of #1657")
670675
@pytest.mark.sklearn()
671676
@pytest.mark.test_server()
672677
def test_run_and_upload_pipeline_dummy_pipeline(self):
@@ -681,6 +686,7 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
681686
n_test_obs = self.TEST_SERVER_TASK_SIMPLE["n_test_obs"]
682687
self._run_and_upload_classification(pipeline1, task_id, n_missing_vals, n_test_obs, "62501")
683688

689+
@pytest.mark.skip(reason="Pending resolution of #1657")
684690
@pytest.mark.sklearn()
685691
@unittest.skipIf(
686692
Version(sklearn.__version__) < Version("0.20"),
@@ -798,6 +804,7 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
798804
call_count += 1
799805
assert call_count == 3
800806

807+
@pytest.mark.skip(reason="Pending resolution of #1657")
801808
@pytest.mark.sklearn()
802809
@pytest.mark.test_server()
803810
def test_run_and_upload_gridsearch(self):
@@ -821,6 +828,7 @@ def test_run_and_upload_gridsearch(self):
821828
)
822829
assert len(run.trace.trace_iterations) == 9
823830

831+
@pytest.mark.skip(reason="Pending resolution of #1657")
824832
@pytest.mark.sklearn()
825833
@pytest.mark.test_server()
826834
def test_run_and_upload_randomsearch(self):
@@ -854,6 +862,7 @@ def test_run_and_upload_randomsearch(self):
854862
trace = openml.runs.get_run_trace(run.run_id)
855863
assert len(trace.trace_iterations) == 5
856864

865+
@pytest.mark.skip(reason="Pending resolution of #1657")
857866
@pytest.mark.sklearn()
858867
@pytest.mark.test_server()
859868
def test_run_and_upload_maskedarrays(self):
@@ -882,6 +891,7 @@ def test_run_and_upload_maskedarrays(self):
882891

883892
##########################################################################
884893

894+
@pytest.mark.skip(reason="Pending resolution of #1657")
885895
@pytest.mark.sklearn()
886896
@pytest.mark.test_server()
887897
def test_learning_curve_task_1(self):
@@ -907,6 +917,7 @@ def test_learning_curve_task_1(self):
907917
)
908918
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
909919

920+
@pytest.mark.skip(reason="Pending resolution of #1657")
910921
@pytest.mark.sklearn()
911922
@pytest.mark.test_server()
912923
def test_learning_curve_task_2(self):
@@ -944,6 +955,7 @@ def test_learning_curve_task_2(self):
944955
)
945956
self._check_sample_evaluations(run.sample_evaluations, num_repeats, num_folds, num_samples)
946957

958+
@pytest.mark.skip(reason="Pending resolution of #1657")
947959
@pytest.mark.sklearn()
948960
@unittest.skipIf(
949961
Version(sklearn.__version__) < Version("0.21"),
@@ -1023,6 +1035,7 @@ def _test_local_evaluations(self, run):
10231035
assert alt_scores[idx] >= 0
10241036
assert alt_scores[idx] <= 1
10251037

1038+
@pytest.mark.skip(reason="Pending resolution of #1657")
10261039
@pytest.mark.sklearn()
10271040
@pytest.mark.test_server()
10281041
def test_local_run_swapped_parameter_order_model(self):
@@ -1039,6 +1052,7 @@ def test_local_run_swapped_parameter_order_model(self):
10391052

10401053
self._test_local_evaluations(run)
10411054

1055+
@pytest.mark.skip(reason="Pending resolution of #1657")
10421056
@pytest.mark.sklearn()
10431057
@unittest.skipIf(
10441058
Version(sklearn.__version__) < Version("0.20"),
@@ -1068,6 +1082,7 @@ def test_local_run_swapped_parameter_order_flow(self):
10681082

10691083
self._test_local_evaluations(run)
10701084

1085+
@pytest.mark.skip(reason="Pending resolution of #1657")
10711086
@pytest.mark.sklearn()
10721087
@unittest.skipIf(
10731088
Version(sklearn.__version__) < Version("0.20"),
@@ -1106,6 +1121,7 @@ def test_online_run_metric_score(self):
11061121

11071122
self._test_local_evaluations(run)
11081123

1124+
@pytest.mark.skip(reason="Pending resolution of #1657")
11091125
@pytest.mark.sklearn()
11101126
@unittest.skipIf(
11111127
Version(sklearn.__version__) < Version("0.20"),
@@ -1168,6 +1184,7 @@ def test_initialize_model_from_run(self):
11681184
assert flowS.components["Imputer"].parameters["strategy"] == '"most_frequent"'
11691185
assert flowS.components["VarianceThreshold"].parameters["threshold"] == "0.05"
11701186

1187+
@pytest.mark.skip(reason="Pending resolution of #1657")
11711188
@pytest.mark.sklearn()
11721189
@unittest.skipIf(
11731190
Version(sklearn.__version__) < Version("0.20"),
@@ -1228,6 +1245,7 @@ def test__run_exists(self):
12281245
run_ids = run_exists(task.task_id, setup_exists)
12291246
assert run_ids, (run_ids, clf)
12301247

1248+
@pytest.mark.skip(reason="Pending resolution of #1657")
12311249
@pytest.mark.sklearn()
12321250
@pytest.mark.test_server()
12331251
def test_run_with_illegal_flow_id(self):
@@ -1248,6 +1266,7 @@ def test_run_with_illegal_flow_id(self):
12481266
avoid_duplicate_runs=True,
12491267
)
12501268

1269+
@pytest.mark.skip(reason="Pending resolution of #1657")
12511270
@pytest.mark.sklearn()
12521271
@pytest.mark.test_server()
12531272
def test_run_with_illegal_flow_id_after_load(self):
@@ -1280,6 +1299,7 @@ def test_run_with_illegal_flow_id_after_load(self):
12801299
TestBase._mark_entity_for_removal("run", loaded_run.run_id)
12811300
TestBase.logger.info(f"collected from test_run_functions: {loaded_run.run_id}")
12821301

1302+
@pytest.mark.skip(reason="Pending resolution of #1657")
12831303
@pytest.mark.sklearn()
12841304
@pytest.mark.test_server()
12851305
def test_run_with_illegal_flow_id_1(self):
@@ -1306,6 +1326,7 @@ def test_run_with_illegal_flow_id_1(self):
13061326
avoid_duplicate_runs=True,
13071327
)
13081328

1329+
@pytest.mark.skip(reason="Pending resolution of #1657")
13091330
@pytest.mark.sklearn()
13101331
@pytest.mark.test_server()
13111332
def test_run_with_illegal_flow_id_1_after_load(self):
@@ -1345,6 +1366,7 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13451366
loaded_run.publish,
13461367
)
13471368

1369+
@pytest.mark.skip(reason="Pending resolution of #1657")
13481370
@pytest.mark.sklearn()
13491371
@unittest.skipIf(
13501372
Version(sklearn.__version__) < Version("0.20"),
@@ -1575,6 +1597,7 @@ def test_get_runs_list_by_tag(self):
15751597
runs = openml.runs.list_runs(tag="curves", size=2)
15761598
assert len(runs) >= 1
15771599

1600+
@pytest.mark.skip(reason="Pending resolution of #1657")
15781601
@pytest.mark.sklearn()
15791602
@unittest.skipIf(
15801603
Version(sklearn.__version__) < Version("0.20"),
@@ -1612,6 +1635,7 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
16121635
# repeat, fold, row_id, 6 confidences, prediction and correct label
16131636
assert len(row) == 12
16141637

1638+
@pytest.mark.skip(reason="Pending resolution of #1657")
16151639
@pytest.mark.sklearn()
16161640
@unittest.skipIf(
16171641
Version(sklearn.__version__) < Version("0.20"),
@@ -1666,6 +1690,7 @@ def test_get_uncached_run(self):
16661690
with pytest.raises(openml.exceptions.OpenMLCacheException):
16671691
openml.runs.functions._get_cached_run(10)
16681692

1693+
@pytest.mark.skip(reason="Pending resolution of #1657")
16691694
@pytest.mark.sklearn()
16701695
@pytest.mark.test_server()
16711696
def test_run_flow_on_task_downloaded_flow(self):
@@ -1767,7 +1792,7 @@ def test_format_prediction_task_regression(self):
17671792
self.assertListEqual(res, [0] * 5)
17681793

17691794

1770-
1795+
@pytest.mark.skip(reason="Pending resolution of #1657")
17711796
@unittest.skipIf(
17721797
Version(sklearn.__version__) < Version("0.20"),
17731798
reason="SimpleImputer doesn't handle mixed type DataFrame as input",
@@ -1864,7 +1889,7 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18641889
assert run_url == mock_delete.call_args.args[0]
18651890
assert test_api_key == mock_delete.call_args.kwargs.get("params", {}).get("api_key")
18661891

1867-
1892+
@pytest.mark.skip(reason="Pending resolution of #1657")
18681893
@pytest.mark.sklearn()
18691894
@unittest.skipIf(
18701895
Version(sklearn.__version__) < Version("0.21"),
@@ -1946,7 +1971,7 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19461971
err_msg="Observed performance scores deviate from expected ones.",
19471972
)
19481973

1949-
1974+
@pytest.mark.skip(reason="Pending resolution of #1657")
19501975
@pytest.mark.sklearn()
19511976
@unittest.skipIf(
19521977
Version(sklearn.__version__) < Version("0.21"),

tests/test_setups/test_setup_functions.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -34,6 +34,7 @@ def setUp(self):
3434
self.extension = SklearnExtension()
3535
super().setUp()
3636

37+
@pytest.mark.skip(reason="Pending resolution of #1657")
3738
@pytest.mark.sklearn()
3839
@pytest.mark.test_server()
3940
def test_nonexisting_setup_exists(self):
@@ -82,6 +83,7 @@ def _existing_setup_exists(self, classif):
8283
setup_id = openml.setups.setup_exists(flow)
8384
assert setup_id == run.setup_id
8485

86+
@pytest.mark.skip(reason="Pending resolution of #1657")
8587
@pytest.mark.sklearn()
8688
@pytest.mark.test_server()
8789
def test_existing_setup_exists_1(self):
@@ -98,12 +100,14 @@ def side_effect(self):
98100
nb = sklearn.naive_bayes.GaussianNB()
99101
self._existing_setup_exists(nb)
100102

103+
@pytest.mark.skip(reason="Pending resolution of #1657")
101104
@pytest.mark.sklearn()
102105
@pytest.mark.test_server()
103106
def test_exisiting_setup_exists_2(self):
104107
# Check a flow with one hyperparameter
105108
self._existing_setup_exists(sklearn.naive_bayes.GaussianNB())
106109

110+
@pytest.mark.skip(reason="Pending resolution of #1657")
107111
@pytest.mark.sklearn()
108112
@pytest.mark.test_server()
109113
def test_existing_setup_exists_3(self):

0 commit comments

Comments
 (0)