@@ -398,6 +398,7 @@ def _check_sample_evaluations(
398398 assert evaluation < max_time_allowed
399399
400400 @pytest .mark .sklearn ()
401+ @pytest .mark .xfail (reason = "failures_issue_1544" )
401402 def test_run_regression_on_classif_task (self ):
402403 task_id = 259 # collins; crossvalidation; has numeric targets
403404
@@ -414,6 +415,7 @@ def test_run_regression_on_classif_task(self):
414415 )
415416
416417 @pytest .mark .sklearn ()
418+ @pytest .mark .xfail (reason = "failures_issue_1544" )
417419 def test_check_erronous_sklearn_flow_fails (self ):
418420 task_id = 115 # diabetes; crossvalidation
419421 task = openml .tasks .get_task (task_id )
@@ -626,6 +628,7 @@ def _run_and_upload_regression(
626628 )
627629
628630 @pytest .mark .sklearn ()
631+ @pytest .mark .xfail (reason = "failures_issue_1544" )
629632 def test_run_and_upload_logistic_regression (self ):
630633 lr = LogisticRegression (solver = "lbfgs" , max_iter = 1000 )
631634 task_id = self .TEST_SERVER_TASK_SIMPLE ["task_id" ]
@@ -634,6 +637,7 @@ def test_run_and_upload_logistic_regression(self):
634637 self ._run_and_upload_classification (lr , task_id , n_missing_vals , n_test_obs , "62501" )
635638
636639 @pytest .mark .sklearn ()
640+ @pytest .mark .xfail (reason = "failures_issue_1544" )
637641 def test_run_and_upload_linear_regression (self ):
638642 lr = LinearRegression ()
639643 task_id = self .TEST_SERVER_TASK_REGRESSION ["task_id" ]
@@ -664,6 +668,7 @@ def test_run_and_upload_linear_regression(self):
664668 self ._run_and_upload_regression (lr , task_id , n_missing_vals , n_test_obs , "62501" )
665669
666670 @pytest .mark .sklearn ()
671+ @pytest .mark .xfail (reason = "failures_issue_1544" )
667672 def test_run_and_upload_pipeline_dummy_pipeline (self ):
668673 pipeline1 = Pipeline (
669674 steps = [
@@ -677,6 +682,7 @@ def test_run_and_upload_pipeline_dummy_pipeline(self):
677682 self ._run_and_upload_classification (pipeline1 , task_id , n_missing_vals , n_test_obs , "62501" )
678683
679684 @pytest .mark .sklearn ()
685+ @pytest .mark .xfail (reason = "failures_issue_1544" )
680686 @unittest .skipIf (
681687 Version (sklearn .__version__ ) < Version ("0.20" ),
682688 reason = "columntransformer introduction in 0.20.0" ,
@@ -793,6 +799,7 @@ def test_run_and_upload_knn_pipeline(self, warnings_mock):
793799 assert call_count == 3
794800
795801 @pytest .mark .sklearn ()
802+ @pytest .mark .xfail (reason = "failures_issue_1544" )
796803 def test_run_and_upload_gridsearch (self ):
797804 estimator_name = (
798805 "base_estimator" if Version (sklearn .__version__ ) < Version ("1.4" ) else "estimator"
@@ -815,6 +822,7 @@ def test_run_and_upload_gridsearch(self):
815822 assert len (run .trace .trace_iterations ) == 9
816823
817824 @pytest .mark .sklearn ()
825+ @pytest .mark .xfail (reason = "failures_issue_1544" )
818826 def test_run_and_upload_randomsearch (self ):
819827 randomsearch = RandomizedSearchCV (
820828 RandomForestClassifier (n_estimators = 5 ),
@@ -847,6 +855,7 @@ def test_run_and_upload_randomsearch(self):
847855 assert len (trace .trace_iterations ) == 5
848856
849857 @pytest .mark .sklearn ()
858+ @pytest .mark .xfail (reason = "failures_issue_1544" )
850859 def test_run_and_upload_maskedarrays (self ):
851860 # This testcase is important for 2 reasons:
852861 # 1) it verifies the correct handling of masked arrays (not all
@@ -874,6 +883,7 @@ def test_run_and_upload_maskedarrays(self):
874883 ##########################################################################
875884
876885 @pytest .mark .sklearn ()
886+ @pytest .mark .xfail (reason = "failures_issue_1544" )
877887 def test_learning_curve_task_1 (self ):
878888 task_id = 801 # diabates dataset
879889 num_test_instances = 6144 # for learning curve
@@ -898,6 +908,7 @@ def test_learning_curve_task_1(self):
898908 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
899909
900910 @pytest .mark .sklearn ()
911+ @pytest .mark .xfail (reason = "failures_issue_1544" )
901912 def test_learning_curve_task_2 (self ):
902913 task_id = 801 # diabates dataset
903914 num_test_instances = 6144 # for learning curve
@@ -934,6 +945,7 @@ def test_learning_curve_task_2(self):
934945 self ._check_sample_evaluations (run .sample_evaluations , num_repeats , num_folds , num_samples )
935946
936947 @pytest .mark .sklearn ()
948+ @pytest .mark .xfail (reason = "failures_issue_1544" )
937949 @unittest .skipIf (
938950 Version (sklearn .__version__ ) < Version ("0.21" ),
939951 reason = "Pipelines don't support indexing (used for the assert check)" ,
@@ -1012,6 +1024,7 @@ def _test_local_evaluations(self, run):
10121024 assert alt_scores [idx ] <= 1
10131025
10141026 @pytest .mark .sklearn ()
1027+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10151028 def test_local_run_swapped_parameter_order_model (self ):
10161029 clf = DecisionTreeClassifier ()
10171030 australian_task = 595 # Australian; crossvalidation
@@ -1027,6 +1040,7 @@ def test_local_run_swapped_parameter_order_model(self):
10271040 self ._test_local_evaluations (run )
10281041
10291042 @pytest .mark .sklearn ()
1043+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10301044 @unittest .skipIf (
10311045 Version (sklearn .__version__ ) < Version ("0.20" ),
10321046 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1055,6 +1069,7 @@ def test_local_run_swapped_parameter_order_flow(self):
10551069 self ._test_local_evaluations (run )
10561070
10571071 @pytest .mark .sklearn ()
1072+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10581073 @unittest .skipIf (
10591074 Version (sklearn .__version__ ) < Version ("0.20" ),
10601075 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1092,6 +1107,7 @@ def test_online_run_metric_score(self):
10921107 self ._test_local_evaluations (run )
10931108
10941109 @pytest .mark .sklearn ()
1110+ @pytest .mark .xfail (reason = "failures_issue_1544" )
10951111 @unittest .skipIf (
10961112 Version (sklearn .__version__ ) < Version ("0.20" ),
10971113 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
@@ -1157,6 +1173,7 @@ def test_initialize_model_from_run(self):
11571173 Version (sklearn .__version__ ) < Version ("0.20" ),
11581174 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
11591175 )
1176+ @pytest .mark .xfail (reason = "failures_issue_1544" )
11601177 def test__run_exists (self ):
11611178 # would be better to not sentinel these clfs,
11621179 # so we do not have to perform the actual runs
@@ -1212,6 +1229,7 @@ def test__run_exists(self):
12121229 assert run_ids , (run_ids , clf )
12131230
12141231 @pytest .mark .sklearn ()
1232+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12151233 def test_run_with_illegal_flow_id (self ):
12161234 # check the case where the user adds an illegal flow id to a
12171235 # non-existing flo
@@ -1231,6 +1249,7 @@ def test_run_with_illegal_flow_id(self):
12311249 )
12321250
12331251 @pytest .mark .sklearn ()
1252+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12341253 def test_run_with_illegal_flow_id_after_load (self ):
12351254 # Same as `test_run_with_illegal_flow_id`, but test this error is also
12361255 # caught if the run is stored to and loaded from disk first.
@@ -1262,6 +1281,7 @@ def test_run_with_illegal_flow_id_after_load(self):
12621281 TestBase .logger .info (f"collected from test_run_functions: { loaded_run .run_id } " )
12631282
12641283 @pytest .mark .sklearn ()
1284+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12651285 def test_run_with_illegal_flow_id_1 (self ):
12661286 # Check the case where the user adds an illegal flow id to an existing
12671287 # flow. Comes to a different value error than the previous test
@@ -1287,6 +1307,7 @@ def test_run_with_illegal_flow_id_1(self):
12871307 )
12881308
12891309 @pytest .mark .sklearn ()
1310+ @pytest .mark .xfail (reason = "failures_issue_1544" )
12901311 def test_run_with_illegal_flow_id_1_after_load (self ):
12911312 # Same as `test_run_with_illegal_flow_id_1`, but test this error is
12921313 # also caught if the run is stored to and loaded from disk first.
@@ -1325,6 +1346,7 @@ def test_run_with_illegal_flow_id_1_after_load(self):
13251346 )
13261347
13271348 @pytest .mark .sklearn ()
1349+ @pytest .mark .xfail (reason = "failures_issue_1544" )
13281350 @unittest .skipIf (
13291351 Version (sklearn .__version__ ) < Version ("0.20" ),
13301352 reason = "OneHotEncoder cannot handle mixed type DataFrame as input" ,
@@ -1552,6 +1574,7 @@ def test_get_runs_list_by_tag(self):
15521574 assert len (runs ) >= 1
15531575
15541576 @pytest .mark .sklearn ()
1577+ @pytest .mark .xfail (reason = "failures_issue_1544" )
15551578 @unittest .skipIf (
15561579 Version (sklearn .__version__ ) < Version ("0.20" ),
15571580 reason = "columntransformer introduction in 0.20.0" ,
@@ -1588,6 +1611,7 @@ def test_run_on_dataset_with_missing_labels_dataframe(self):
15881611 assert len (row ) == 12
15891612
15901613 @pytest .mark .sklearn ()
1614+ @pytest .mark .xfail (reason = "failures_issue_1544" )
15911615 @unittest .skipIf (
15921616 Version (sklearn .__version__ ) < Version ("0.20" ),
15931617 reason = "columntransformer introduction in 0.20.0" ,
@@ -1640,6 +1664,7 @@ def test_get_uncached_run(self):
16401664 openml .runs .functions ._get_cached_run (10 )
16411665
16421666 @pytest .mark .sklearn ()
1667+ @pytest .mark .xfail (reason = "failures_issue_1544" )
16431668 def test_run_flow_on_task_downloaded_flow (self ):
16441669 model = sklearn .ensemble .RandomForestClassifier (n_estimators = 33 )
16451670 flow = self .extension .model_to_flow (model )
@@ -1740,6 +1765,7 @@ def test_format_prediction_task_regression(self):
17401765 reason = "SimpleImputer doesn't handle mixed type DataFrame as input" ,
17411766 )
17421767 @pytest .mark .sklearn ()
1768+ @pytest .mark .xfail (reason = "failures_issue_1544" )
17431769 def test_delete_run (self ):
17441770 rs = np .random .randint (1 , 2 ** 31 - 1 )
17451771 clf = sklearn .pipeline .Pipeline (
@@ -1835,6 +1861,7 @@ def test_delete_unknown_run(mock_delete, test_files_directory, test_api_key):
18351861
18361862
18371863@pytest .mark .sklearn ()
1864+ @pytest .mark .xfail (reason = "failures_issue_1544" )
18381865@unittest .skipIf (
18391866 Version (sklearn .__version__ ) < Version ("0.21" ),
18401867 reason = "couldn't perform local tests successfully w/o bloating RAM" ,
@@ -1930,6 +1957,7 @@ def test__run_task_get_arffcontent_2(parallel_mock):
19301957 (- 1 , "threading" , 10 ), # the threading backend does preserve mocks even with parallelizing
19311958 ]
19321959)
1960+ @pytest .mark .xfail (reason = "failures_issue_1544" )
19331961def test_joblib_backends (parallel_mock , n_jobs , backend , call_count ):
19341962 """Tests evaluation of a run using various joblib backends and n_jobs."""
19351963 if backend is None :
0 commit comments