pred_time_limit clarification and logging (#319)

* pred_time_limit clarification

* log prediction time

* handle ChunkedEncodingError in test
This commit is contained in:
Chi Wang
2021-12-03 16:02:00 -08:00
committed by GitHub
parent fb59bb9928
commit 18230ed22f
3 changed files with 8 additions and 5 deletions

View File

@@ -1,4 +1,5 @@
from openml.exceptions import OpenMLServerException
from requests.exceptions import ChunkedEncodingError
def test_automl(budget=5, dataset_format="dataframe", hpo_method=None):
@@ -8,8 +9,8 @@ def test_automl(budget=5, dataset_format="dataframe", hpo_method=None):
X_train, X_test, y_train, y_test = load_openml_dataset(
dataset_id=1169, data_dir="test/", dataset_format=dataset_format
)
except OpenMLServerException:
print("OpenMLServerException raised")
except (OpenMLServerException, ChunkedEncodingError) as e:
print(e)
return
""" import AutoML class from flaml package """
from flaml import AutoML
@@ -84,8 +85,8 @@ def test_mlflow():
X_train, X_test, y_train, y_test = load_openml_task(
task_id=7592, data_dir="test/"
)
except OpenMLServerException:
print("OpenMLServerException raised")
except (OpenMLServerException, ChunkedEncodingError) as e:
print(e)
return
""" import AutoML class from flaml package """
from flaml import AutoML