Source code for gemseo.mlearning.quality_measures.error_measure

# Copyright 2021 IRT Saint Exupéry, https://www.irt-saintexupery.com
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License version 3 as published by the Free Software Foundation.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public License
# along with this program; if not, write to the Free Software Foundation,
# Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
# Contributors:
#    INITIAL AUTHORS - initial API and implementation and/or initial
#                         documentation
#        :author: Syver Doving Agdestein
#    OTHER AUTHORS   - MACROSCOPIC CHANGES
"""Here is the baseclass to measure the error of machine learning algorithms.

The concept of error measure is implemented with the :class:`.MLErrorMeasure` class and
proposes different evaluation methods.
"""

from __future__ import annotations

from abc import abstractmethod
from typing import TYPE_CHECKING
from typing import Final

from numpy import atleast_1d
from numpy import ndarray

from gemseo.mlearning.quality_measures.quality_measure import MeasureType
from gemseo.mlearning.quality_measures.quality_measure import MLQualityMeasure
from gemseo.mlearning.resampling.bootstrap import Bootstrap
from gemseo.mlearning.resampling.cross_validation import CrossValidation
from gemseo.utils.data_conversion import split_array_to_dict_of_arrays

if TYPE_CHECKING:
    from collections.abc import Sequence

    from gemseo.datasets.io_dataset import IODataset
    from gemseo.mlearning.core.supervised import MLSupervisedAlgo


[docs] class MLErrorMeasure(MLQualityMeasure): """An abstract error measure for machine learning.""" __OUTPUT_NAME_SEPARATOR: Final[str] = "#" """A string to join output names.""" _GEMSEO_MULTIOUTPUT_TO_SKLEARN_MULTIOUTPUT: Final[dict[bool, str]] = { True: "raw_values", False: "uniform_average", } """Map from the argument "multioutput" of |g| to that of sklearn.""" algo: MLSupervisedAlgo def __init__( self, algo: MLSupervisedAlgo, fit_transformers: bool = MLQualityMeasure._FIT_TRANSFORMERS, ) -> None: """ Args: algo: A machine learning algorithm for supervised learning. """ # noqa: D205 D212 super().__init__(algo, fit_transformers=fit_transformers)
[docs] def compute_learning_measure( self, samples: Sequence[int] | None = None, multioutput: bool = True, as_dict: bool = False, ) -> MeasureType: """ Args: as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. """ # noqa: D205 D212 self._pre_process(samples) return self._post_process_measure( self._compute_measure( self.algo.output_data, self.algo.predict(self.algo.input_data), multioutput, ), multioutput, as_dict, )
[docs] def compute_test_measure( self, test_data: IODataset, samples: Sequence[int] | None = None, multioutput: bool = True, as_dict: bool = False, ) -> MeasureType: """ Args: as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. """ # noqa: D205 D212 self._pre_process(samples) return self._post_process_measure( self._compute_measure( test_data.get_view( group_names=test_data.OUTPUT_GROUP, variable_names=self.algo.output_names, ).to_numpy(), self.algo.predict( test_data.get_view( group_names=test_data.INPUT_GROUP, variable_names=self.algo.input_names, ).to_numpy() ), multioutput, ), multioutput, as_dict, )
[docs] def compute_leave_one_out_measure( self, samples: Sequence[int] | None = None, multioutput: bool = True, as_dict: bool = False, store_resampling_result: bool = False, ) -> MeasureType: """ Args: as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. """ # noqa: D205 D212 return self.compute_cross_validation_measure( samples=samples, n_folds=self.algo.learning_set.n_samples, multioutput=multioutput, as_dict=as_dict, store_resampling_result=store_resampling_result, seed=1, )
[docs] def compute_cross_validation_measure( self, n_folds: int = 5, samples: Sequence[int] | None = None, multioutput: bool = True, randomize: bool = MLQualityMeasure._RANDOMIZE, seed: int | None = None, as_dict: bool = False, store_resampling_result: bool = False, ) -> MeasureType: """ Args: as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. """ # noqa: D205 D212 samples, seed = self._pre_process(samples, seed, randomize) cross_validation = CrossValidation(samples, n_folds, randomize, seed) output_data = self.algo.output_data _, predictions = cross_validation.execute( self.algo, store_resampling_result, True, True, self._fit_transformers, store_resampling_result, self.algo.input_data, output_data.shape, ) return self._post_process_measure( self._compute_measure(output_data, predictions, multioutput), multioutput, as_dict, )
[docs] def compute_bootstrap_measure( self, n_replicates: int = 100, samples: Sequence[int] | None = None, multioutput: bool = True, seed: None | None = None, as_dict: bool = False, store_resampling_result: bool = False, ) -> MeasureType: """ Args: as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. """ # noqa: D205 D212 samples, seed = self._pre_process(samples, seed, True) bootstrap = Bootstrap(samples, n_replicates, seed) output_data = self.algo.output_data _, predictions = bootstrap.execute( self.algo, store_resampling_result, True, False, self._fit_transformers, store_resampling_result, self.algo.input_data, output_data.shape, ) measure = 0 for prediction, split in zip(predictions, bootstrap.splits): measure += self._compute_measure( output_data[split.test], prediction, multioutput ) return self._post_process_measure(measure / n_replicates, multioutput, as_dict)
@abstractmethod def _compute_measure( self, outputs: ndarray, predictions: ndarray, multioutput: bool = True, ) -> MeasureType: """Compute the quality measure. Args: outputs: The reference data. predictions: The predicted labels. multioutput: Whether the quality measure is returned for each component of the outputs. Otherwise, the average quality measure. Returns: The value of the quality measure. """ def _post_process_measure( self, measure: float | ndarray, multioutput: bool, as_dict: bool ) -> MeasureType: """Post-process a measure. Args: measure: The measure to post-process. multioutput: Whether the quality measure is returned for each component of the outputs. Otherwise, the average quality measure. as_dict: Whether the full quality measure is returned as a mapping from ``algo.output names`` to quality measures. Otherwise, the full quality measure as an array stacking these quality measures according to the order of ``algo.output_names``. Returns: The post-processed measure. """ if not as_dict: return measure data = atleast_1d(measure) names = self.algo.output_names if not multioutput: return {self.__OUTPUT_NAME_SEPARATOR.join(names): data} return split_array_to_dict_of_arrays( data, self.algo.learning_set.variable_names_to_n_components, names ) # TODO: API: remove these aliases in the next major release. evaluate_learn = compute_learning_measure evaluate_test = compute_test_measure evaluate_kfolds = compute_cross_validation_measure evaluate_loo = compute_leave_one_out_measure evaluate_bootstrap = compute_bootstrap_measure