Skip to content

Commit

Permalink
Fix MASE error by passing through y_train (#4258)
Browse files Browse the repository at this point in the history
* Add y_train as an argument to score_all_objectives
  • Loading branch information
eccabay authored Aug 4, 2023
1 parent 30f3242 commit 045686d
Show file tree
Hide file tree
Showing 6 changed files with 66 additions and 6 deletions.
1 change: 1 addition & 0 deletions docs/source/release_notes.rst
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ Release Notes
* Added stacking and unstacking utility functions to work with multiseries data :pr:`4250`
* Fixes
* Added support for pandas 2 :pr:`4216`
* Fixed bug where time series pipelines would fail due to MASE needing `y_train` when scoring :pr:`4258`
* Update s3 bucket for docs image :pr:`4260`
* Changes
* Unpinned sktime version :pr:`4214`
Expand Down
2 changes: 1 addition & 1 deletion evalml/pipelines/binary_classification_pipeline.py
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ def predict_proba(self, X, X_train=None, y_train=None):
return super().predict_proba(X)

@staticmethod
def _score(X, y, predictions, objective):
def _score(X, y, predictions, objective, y_train=None):
"""Given data, model predictions or predicted probabilities computed on the data, and an objective, evaluate and return the objective score."""
if predictions.ndim > 1:
predictions = predictions.iloc[:, 1]
Expand Down
16 changes: 13 additions & 3 deletions evalml/pipelines/pipeline_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -351,10 +351,18 @@ def score(self, X, y, objectives, X_train=None, y_train=None):
"""

@staticmethod
def _score(X, y, predictions, objective):
return objective.score(y, predictions, X=X)
def _score(X, y, predictions, objective, y_train=None):
return objective.score(y, predictions, X=X, y_train=y_train)

def _score_all_objectives(self, X, y, y_pred, y_pred_proba, objectives):
def _score_all_objectives(
self,
X,
y,
y_pred,
y_pred_proba,
objectives,
y_train=None,
):
"""Given data, model predictions or predicted probabilities computed on the data, and an objective, evaluate and return the objective score.
Will raise a PipelineScoreError if any objectives fail.
Expand All @@ -366,6 +374,7 @@ def _score_all_objectives(self, X, y, y_pred, y_pred_proba, objectives):
y_pred_proba (pd.Dataframe, pd.Series, None): The predicted probabilities for classification problems.
Will be a DataFrame for multiclass problems and Series otherwise. Will be None for regression problems.
objectives (list): List of objectives to score.
y_train (pd.Series or None): Training labels. Only used for time series, otherwise ignored.
Returns:
dict: Ordered dictionary with objectives and their scores.
Expand All @@ -390,6 +399,7 @@ def _score_all_objectives(self, X, y, y_pred, y_pred_proba, objectives):
y,
y_pred_proba if objective.score_needs_proba else y_pred,
objective,
y_train,
)
scored_successfully.update({objective.name: score})
except Exception as e:
Expand Down
10 changes: 8 additions & 2 deletions evalml/pipelines/time_series_classification_pipelines.py
Original file line number Diff line number Diff line change
Expand Up @@ -282,11 +282,17 @@ def predict_in_sample(self, X, y, X_train, y_train, objective=None):
return infer_feature_types(predictions)

@staticmethod
def _score(X, y, predictions, objective):
def _score(X, y, predictions, objective, y_train=None):
"""Given data, model predictions or predicted probabilities computed on the data, and an objective, evaluate and return the objective score."""
if predictions.ndim > 1:
predictions = predictions.iloc[:, 1]
return TimeSeriesClassificationPipeline._score(X, y, predictions, objective)
return TimeSeriesClassificationPipeline._score(
X,
y,
predictions,
objective,
y_train,
)


class TimeSeriesMulticlassClassificationPipeline(TimeSeriesClassificationPipeline):
Expand Down
1 change: 1 addition & 0 deletions evalml/pipelines/time_series_regression_pipeline.py
Original file line number Diff line number Diff line change
Expand Up @@ -102,6 +102,7 @@ def score(self, X, y, objectives, X_train=None, y_train=None):
y_predicted,
y_pred_proba=None,
objectives=objectives,
y_train=y_train,
)

def get_forecast_period(self, X):
Expand Down
42 changes: 42 additions & 0 deletions evalml/tests/pipeline_tests/test_pipelines.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
import io
import math
import os
import pickle
import re
Expand Down Expand Up @@ -1048,6 +1049,47 @@ def test_score_with_objective_that_requires_predict_proba(
mock_predict.assert_called()


@patch("evalml.pipelines.components.Estimator.predict")
@patch("evalml.pipelines.components.Estimator.fit")
def test_score_with_objective_that_requires_y_train(
mock_fit,
mock_predict,
dummy_time_series_regression_pipeline_class,
generate_seasonal_data,
):
X, y = generate_seasonal_data(real_or_synthetic="real")(period=10)
X = X.reset_index()

split = math.floor(0.9 * len(X))
X_train, X_holdout = X.iloc[:split], X.iloc[split:]
y_train, y_holdout = y.iloc[:split], y.iloc[split:]

parameters = {
"pipeline": {
"max_delay": 0,
"gap": 2,
"forecast_horizon": 2,
"time_index": "Date",
},
}

mock_regression_pipeline = dummy_time_series_regression_pipeline_class(
parameters=parameters,
)

mock_predict.return_value = pd.Series([1] * len(y_holdout))

mock_regression_pipeline.fit(X_train, y_train)
mock_regression_pipeline.score(
X_holdout,
y_holdout,
["mean absolute scaled error"],
X_train=X_train,
y_train=y_train,
)
mock_predict.assert_called()


def test_score_auc(X_y_binary, logistic_regression_binary_pipeline):
X, y = X_y_binary
lr_pipeline = logistic_regression_binary_pipeline
Expand Down

0 comments on commit 045686d

Please sign in to comment.