@@ -413,6 +413,7 @@ def test_run_regression_on_classif_task(self):
413413 task = task ,
414414 )
415415
416+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
416417 @pytest .mark .sklearn ()
417418 def test_check_erronous_sklearn_flow_fails (self ):
418419 task_id = 115 # diabetes; crossvalidation
@@ -881,6 +882,7 @@ def test_run_and_upload_maskedarrays(self):
881882
882883 ##########################################################################
883884
885+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
884886 @pytest .mark .sklearn ()
885887 def test_learning_curve_task_1 (self ):
886888 task_id = 801 # diabates dataset
@@ -905,6 +907,7 @@ def test_learning_curve_task_1(self):
905907 )
906908 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
907909
910+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
908911 @pytest .mark .sklearn ()
909912 def test_learning_curve_task_2 (self ):
910913 task_id = 801 # diabates dataset
@@ -941,6 +944,7 @@ def test_learning_curve_task_2(self):
941944 )
942945 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
943946
947+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
944948 @pytest .mark .sklearn ()
945949 @unittest .skipIf (
946950 Version (sklearn .__version__ ) < Version ("0.21" ),
@@ -1019,6 +1023,7 @@ def _test_local_evaluations(self, run):
10191023 assert alt_scores [idx ] >= 0
10201024 assert alt_scores [idx ] <= 1
10211025
1026+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
10221027 @pytest .mark .sklearn ()
10231028 def test_local_run_swapped_parameter_order_model (self ):
10241029 clf = DecisionTreeClassifier ()
@@ -1034,6 +1039,7 @@ def test_local_run_swapped_parameter_order_model(self):
10341039
10351040 self ._test_local_evaluations (run )
10361041
1042+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
10371043 @pytest .mark .sklearn ()
10381044 @unittest .skipIf (
10391045 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1062,6 +1068,7 @@ def test_local_run_swapped_parameter_order_flow(self):
10621068
10631069 self ._test_local_evaluations (run )
10641070
1071+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
10651072 @pytest .mark .sklearn ()
10661073 @unittest .skipIf (
10671074 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1099,6 +1106,7 @@ def test_online_run_metric_score(self):
10991106
11001107 self ._test_local_evaluations (run )
11011108
1109+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
11021110 @pytest .mark .sklearn ()
11031111 @unittest .skipIf (
11041112 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1160,6 +1168,7 @@ def test_initialize_model_from_run(self):
11601168 assert flowS .components ["Imputer" ].parameters ["strategy" ] == '"most_frequent"'
11611169 assert flowS .components ["VarianceThreshold" ].parameters ["threshold" ] == "0.05"
11621170
1171+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
11631172 @pytest .mark .sklearn ()
11641173 @unittest .skipIf (
11651174 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1219,6 +1228,7 @@ def test__run_exists(self):
12191228 run_ids = run_exists (task .task_id , setup_exists )
12201229 assert run_ids , (run_ids , clf )
12211230
1231+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
12221232 @pytest .mark .sklearn ()
12231233 def test_run_with_illegal_flow_id (self ):
12241234 # check the case where the user adds an illegal flow id to a
@@ -1238,6 +1248,7 @@ def test_run_with_illegal_flow_id(self):
12381248 avoid_duplicate_runs = True ,
12391249 )
12401250
1251+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
12411252 @pytest .mark .sklearn ()
12421253 def test_run_with_illegal_flow_id_after_load (self ):
12431254 # Same as `test_run_with_illegal_flow_id`, but test this error is also
@@ -1294,6 +1305,7 @@ def test_run_with_illegal_flow_id_1(self):
12941305 avoid_duplicate_runs = True ,
12951306 )
12961307
1308+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
12971309 @pytest .mark .sklearn ()
12981310 def test_run_with_illegal_flow_id_1_after_load (self ):
12991311 # Same as `test_run_with_illegal_flow_id_1`, but test this error is
@@ -1332,6 +1344,7 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13321344 loaded_run .publish ,
13331345 )
13341346
1347+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
13351348 @pytest .mark .sklearn ()
13361349 @unittest .skipIf (
13371350 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1559,6 +1572,7 @@ def test_get_runs_list_by_tag(self):
15591572 runs = openml .runs .list_runs (tag = "curves" , size = 2 )
15601573 assert len (runs ) >= 1
15611574
1575+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
15621576 @pytest .mark .sklearn ()
15631577 @unittest .skipIf (
15641578 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1595,6 +1609,7 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
15951609 # repeat, fold, row_id, 6 confidences, prediction and correct label
15961610 assert len (row ) == 12
15971611
1612+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
15981613 @pytest .mark .sklearn ()
15991614 @unittest .skipIf (
16001615 Version (sklearn .__version__ ) < Version ("0.20" ),
@@ -1647,6 +1662,7 @@ def test_get_uncached_run(self):
16471662 with pytest .raises (openml .exceptions .OpenMLCacheException ):
16481663 openml .runs .functions ._get_cached_run (10 )
16491664
1665+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
16501666 @pytest .mark .sklearn ()
16511667 def test_run_flow_on_task_downloaded_flow (self ):
16521668 model = sklearn .ensemble .RandomForestClassifier (n_estimators = 33 )
@@ -1687,6 +1703,7 @@ def test_format_prediction_classification_no_probabilities(self):
16871703 with pytest .raises (ValueError , match = "`proba` is required for classification task" ):
16881704 format_prediction (classification , * ignored_input , proba = None )
16891705
1706+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
16901707 def test_format_prediction_classification_incomplete_probabilities (self ):
16911708 classification = openml .tasks .get_task (
16921709 self .TEST_SERVER_TASK_SIMPLE ["task_id" ],
@@ -1707,13 +1724,15 @@ def test_format_prediction_task_without_classlabels_set(self):
17071724 with pytest .raises (ValueError , match = "The classification task must have class labels set" ):
17081725 format_prediction (classification , * ignored_input , proba = {})
17091726
1727+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
17101728 def test_format_prediction_task_learning_curve_sample_not_set (self ):
17111729 learning_curve = openml .tasks .get_task (801 , download_data = False ) # diabetes;crossvalidation
17121730 probabilities = {c : 0.2 for c in learning_curve .class_labels }
17131731 ignored_input = [0 ] * 5
17141732 with pytest .raises (ValueError , match = "`sample` can not be none for LearningCurveTask" ):
17151733 format_prediction (learning_curve , * ignored_input , sample = None , proba = probabilities )
17161734
1735+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
17171736 def test_format_prediction_task_regression (self ):
17181737 task_meta_data = self .TEST_SERVER_TASK_REGRESSION ["task_meta_data" ]
17191738 _task_id = check_task_existence (** task_meta_data )
@@ -1743,6 +1762,7 @@ def test_format_prediction_task_regression(self):
17431762
17441763
17451764
1765+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
17461766 @unittest .skipIf (
17471767 Version (sklearn .__version__ ) < Version ("0.20" ),
17481768 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1843,6 +1863,7 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18431863
18441864
18451865@pytest .mark .sklearn ()
1866+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
18461867@unittest .skipIf (
18471868 Version (sklearn .__version__ ) < Version ("0.21" ),
18481869 reason = "couldn't perform local tests successfully w/o bloating RAM" ,
@@ -1919,6 +1940,7 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19191940 )
19201941
19211942
1943+ @pytest .mark .xfail (reason = "failures_issue_1544" , strict = False )
19221944@pytest .mark .sklearn ()
19231945@unittest .skipIf (
19241946 Version (sklearn .__version__ ) < Version ("0.21" ),
0 commit comments