Source code for zfit.minimizers.fitresult

#  Copyright (c) 2024 zfit

from __future__ import annotations

import typing
from typing import TYPE_CHECKING

from ..util.exception import BreakingAPIChangeError

if TYPE_CHECKING:
    import zfit

    from .evaluation import LossEval

import collections
import contextlib
import itertools
import math
import warnings
from collections.abc import Callable, Iterable, Mapping

import colored
import iminuit
import numpy as np
import scipy.optimize
import scipy.stats
from colorama import Style, init
from ordered_set import OrderedSet
from scipy.optimize import LbfgsInvHessProduct
from tabulate import tabulate

if TYPE_CHECKING:
    with contextlib.suppress(ImportError):  # for type checking
        import ipyopt

from ..core.interfaces import (
    ZfitData,
    ZfitIndependentParameter,
    ZfitLoss,
    ZfitParameter,
)
from ..core.parameter import set_values
from ..util.container import convert_to_container
from ..util.deprecation import deprecated, deprecated_args
from ..util.warnings import ExperimentalFeatureWarning
from ..util.ztyping import ParamsTypeOpt
from ..z import numpy as znp
from .errors import (
    compute_errors,
    covariance_with_weights,
    dict_to_matrix,
    matrix_to_dict,
)
from .interface import ZfitMinimizer, ZfitResult
from .termination import ConvergenceCriterion

init(autoreset=True)


class Approximations:
    def __init__(
        self,
        params: list[ZfitParameter],
        gradient: np.ndarray | None = None,
        hessian: np.ndarray | None = None,
        inv_hessian: np.ndarray | None = None,
    ) -> None:
        """Holds different approximations after the minimisation and/or calculates them.

        Args:
            params: List of parameters the approximations (gradient, hessian, ...) were calculated with.
            gradient: Gradient
            hessian: Hessian Matrix
            inv_hessian: Inverse of the Hessian Matrix
        """
        self._params = params
        self._gradient = gradient
        self._hessian = hessian
        self._inv_hessian = inv_hessian
        super().__init__()

    @property
    def params(self):
        return self._params

    def gradient(self, params: ZfitParameter | Iterable[ZfitParameter] | None = None) -> np.ndarray | None:
        """Return an approximation of the gradient _if available_.

        Args:
            params: Parameters to which the gradients should be returned

        Returns:
            Array with gradients or ``None``
        """
        grad = self._gradient
        if grad is None:
            return None

        if params is not None:
            params = convert_to_container(params, container=tuple)
            params_mapped = {i: params.index(param) for i, param in enumerate(self.params) if param in params}
            indices = sorted(params_mapped, key=lambda x: params_mapped[x])
            grad = grad[indices]
        return grad

    def hessian(self, invert: bool = True) -> np.ndarray | None:
        """Return an approximation of the hessian _if available_.

        Args:
            invert: If a _hessian approximation_ is not available but an inverse hessian is, invert the latter to
                obtain the hessian approximation.

        Returns:
            Array with hessian matrix or ``None``
        """
        hess = self._hessian
        if hess is None and invert:
            inv_hess = self._inv_hessian
            if inv_hess is not None:
                hess = np.linalg.inv(inv_hess)
                self._hessian = hess
        return hess

    def inv_hessian(self, invert: bool = True) -> None | np.ndarray:
        """Return an approximation of the inverse hessian _if available_.

        Args:
            invert: If an _inverse hessian approximation_ is not available but a hessian is, invert the latter to
                obtain the inverse hessian approximation.

        Returns:
            Array with the inverse of the hessian matrix or ``None``
        """
        inv_hess = self._inv_hessian
        if inv_hess is None and invert:
            hess = self._hessian
            if hess is not None:
                inv_hess = np.linalg.inv(hess)
                self._inv_hessian = inv_hess
        return inv_hess

    def freeze(self):
        self._params = [p.name for p in self.params]


def _minos_minuit(result, params, cl=None):
    minuit_minimizer = result._create_minuit_instance()

    try:
        minuit_minimizer.minos(*(p.name for p in params), cl=cl)
        # Minuit seems very ustable on this and the call can fail after a few trials
    except RuntimeError as error:
        if "Function minimum is not valid." not in error.args[0]:
            raise
        minuit_minimizer.reset()
        minuit_minimizer.minos(*(p.name for p in params), cl=cl)

    merror_result = minuit_minimizer.merrors  # returns every var
    attrs = [
        "lower",
        "upper",
        "is_valid",
        "upper_valid",
        "lower_valid",
        "at_lower_limit",
        "at_upper_limit",
        "nfcn",
    ]
    errors = {}
    for p in params:
        error_res = merror_result[p.name]
        errors[p] = {attr: getattr(error_res, attr) for attr in attrs}
        errors[p]["original"] = error_res
    new_result = None
    return errors, new_result


def _covariance_minuit(result, params):
    minuit = result._create_minuit_instance()

    _ = minuit.hesse()  # make sure to have an accurate covariance

    covariance = minuit.covariance

    covariance_dict = {}
    if covariance is None:
        warnings.warn(
            "minuit failed to calculate the covariance matrix or similar when calling `hesse`."
            "Try to use `hesse_np` as the method instead and try again."
            "This is unexpected and may has to do with iminuitV2. Either way, please fill an issue if"
            " this is not expected to fail for you.",
            RuntimeWarning,
            stacklevel=2,
        )
    else:
        for p1 in params:
            for p2 in params:
                key = (p1, p2)
                covariance_dict[key] = covariance[tuple(k.name for k in key)]

    return covariance_dict


def _covariance_np(result, params):
    if any(isinstance(data, ZfitData) and data.weights is not None for data in result.loss.data):
        warnings.warn(
            "The computation of the covariance matrix with weights is still experimental.",
            ExperimentalFeatureWarning,
            stacklevel=3,
        )

    _, gradient, hessian = result.loss.value_gradient_hessian(params)
    covariance = znp.linalg.inv(hessian)

    return matrix_to_dict(params, covariance)


def _covariance_approx(result, params):
    if any(isinstance(data, ZfitData) and data.weights is not None for data in result.loss.data):
        warnings.warn(
            "Approximate covariance/hesse estimation with weights is not supported, returning None",
            RuntimeWarning,
            stacklevel=3,
        )

    inv_hessian = result.approx.inv_hessian(invert=True)
    if inv_hessian is None:
        return {}

    params_approx = list(result.params)
    param_indices = [params_approx.index(p) for p in params]
    return {
        (p1, p2): inv_hessian[(p1_index, p2_index)]
        for (p1, p1_index), (p2, p2_index) in itertools.product(zip(params, param_indices), zip(params, param_indices))
    }


class ParamToNameGetitem:
    __slots__ = ()

    def __getitem__(self, item):
        if isinstance(item, ZfitParameter):
            item = item.name
        return super().__getitem__(item)


class NameToParamGetitem:
    __slots__ = ()

    def __getitem__(self, item):
        if isinstance(item, ZfitParameter):
            item = item.name
        for param in self.keys():
            name = param.name if isinstance(param, ZfitParameter) else param
            if name == item:
                item = param
                break
        return super().__getitem__(item)  # raises key error if not there, which is good

    def __contains__(self, item):
        try:
            self[item]
        except KeyError:
            return False
        except Exception as error:
            msg = "Unknown exception occurred! This should not happen."
            raise RuntimeError(msg) from error
        else:
            return True


[docs] class FitResult(ZfitResult): _default_hesse = "minuit_hesse" _hesse_methods: typing.ClassVar = { "minuit_hesse": _covariance_minuit, "hesse_np": _covariance_np, "approx": _covariance_approx, } _default_error = "minuit_minos" _error_methods: typing.ClassVar = { "minuit_minos": _minos_minuit, "zfit_error": compute_errors, "zfit_errors": compute_errors, } def __init__( self, loss: ZfitLoss, params: dict[ZfitParameter, float], minimizer: ZfitMinimizer, valid: bool, edm: float, fminopt: float, criterion: ConvergenceCriterion | None, status: int | None = None, converged: bool | None = None, message: str | None = None, info: Mapping | None = None, approx: Mapping | Approximations | None = None, niter: int | None = None, evaluator: LossEval = None, ) -> None: """Create a ``FitResult`` from a minimization. Store parameter values, minimization infos and calculate errors. Any errors calculated are saved under ``self.params`` dictionary with:: {parameter: {error_name1: {'low': value, 'high': value or similar}} The ``FitResult can be used to temporarily update the parameter values to the values found by the minimization .. code-block:: with result: # do something with the new values ... Args: loss: |@doc:result.init.loss| The loss function that was minimized. Usually, but not necessary, contains also the pdf, data and constraints. |@docend:result.init.loss| params: |@doc:result.init.params| Result of the fit where each :py:class:`~zfit.Parameter` key has the value from the minimum found by the minimizer. |@docend:result.init.params| minimizer: |@doc:result.init.minimizer| Minimizer that was used to obtain this ``FitResult`` and will be used to calculate certain errors. If the minimizer is state-based (like "iminuit"), then this is a copy and the state of other ``FitResults`` or of the *actual* minimizer that performed the minimization won't be altered. |@docend:result.init.minimizer| valid: |@doc:result.init.valid| Indicating whether the result is valid or not. This is the strongest indication and serves as the global flag. The reasons why a result may be invalid can be arbitrary, including but not exclusive: - parameter(s) at the limit - maxiter reached without proper convergence - the minimizer maybe even converged but it is known that this is only a local minimum To indicate the reason for the invalidity, pass a message. |@docend:result.init.valid| edm: |@doc:result.init.edm| The estimated distance to minimum which is the criterion value at the minimum. |@docend:result.init.edm| fminopt: |@doc:result.init.fmin| Value of the function at the minimum. |@docend:result.init.fmin| criterion: |@doc:result.init.criterion| Criterion that was used during the minimization. This determines the estimated distance to the minimum (edm) |@docend:result.init.criterion| status: |@doc:result.init.status| A status code (if available) that describes the minimization termination. 0 means a valid termination. |@docend:result.init.status| converged: |@doc:result.init.converged| Whether the fit has successfully converged or not. The result itself can still be an invalid minimum such as if the parameters are at or close to the limits or in case another minimum is found. |@docend:result.init.converged| message: |@doc:result.init.message| Human-readable message to indicate the reason if the fitresult is not valid. If the fit is valid, the message (should) be an empty string (or None), otherwise, it should denote the reason for the invalidity. |@docend:result.init.message| info: |@doc:result.init.info| Additional information (if available) such as *number of gradient function calls* or the original minimizer return message. This is a relatively free field and _no single field_ in it is guaranteed to be stable. Some recommended fields: - *original*: contains the original returned object by the minimizer used internally. - *optimizer*: the actual instance of the wrapped optimizer (if available) |@docend:result.init.info| approx: |@doc:result.init.approx| Collection of approximations found during the minimization process such as gradient and hessian. |@docend:result.init.approx| niter: |@doc:result.init.niter| Approximate number of iterations ~= number of function evaluations ~= number of gradient evaluations. This is an approximated value and the exact meaning can differ between different minimizers. |@docend:result.init.niter| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| """ super().__init__() if status is None: status = 0 if valid else -999 if converged is None and valid: converged = True if message is None: message = "" if valid else "Invalid, unknown reason (not specified)" info = {} if info is None else info approx = self._input_convert_approx(approx, evaluator, info, params) if evaluator is not None: niter = evaluator.niter if niter is None else niter param_at_limit = any(param.at_limit for param in params) if param_at_limit: valid = False if message: message += " AND " message += "parameter(s) at their limit." self._cache_minuit = None # in case used in errors self._evaluator = evaluator # keep private for now self._niter = niter # keep private for now self._approx = approx self._status = status self._message = "" if message is None else message self._converged = converged self._params = self._input_convert_params(params) self._values = ValuesHolder(params) self._params_at_limit = param_at_limit self._edm = edm self._criterion = criterion self._fminopt = fminopt self._info = info self._loss = loss self._minimizer = minimizer self._valid = valid self._covariance_dict = {} self._tmp_old_param_values = [] try: fminfull = loss.value(full=True) except Exception as error: warnings.warn( f"Could not calculate fminfull due to {error}. Setting to 0. This is a new feature and is caught to not break backwards compatibility.", RuntimeWarning, stacklevel=3, ) fminfull = 0 self._fmin = float(fminfull) def _input_convert_approx(self, approx, evaluator, info, params): """Convert approx (if a Mapping) to an `Approximation` using the information provided. Args: approx: |@doc:result.init.approx| Collection of approximations found during the minimization process such as gradient and hessian. |@docend:result.init.approx| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| info: |@doc:result.init.info| Additional information (if available) such as *number of gradient function calls* or the original minimizer return message. This is a relatively free field and _no single field_ in it is guaranteed to be stable. Some recommended fields: - *original*: contains the original returned object by the minimizer used internally. - *optimizer*: the actual instance of the wrapped optimizer (if available) |@docend:result.init.info| params: |@doc:result.init.params| Result of the fit where each :py:class:`~zfit.Parameter` key has the value from the minimum found by the minimizer. |@docend:result.init.params| Returns: The created approximation. """ approx = {} if approx is None else approx if isinstance(approx, collections.abc.Mapping): if "params" not in approx: approx["params"] = params if info: if "gradient" not in approx: approx["gradient"] = info.get("grad", info.get("gradient")) if "hessian" not in approx: approx["hessian"] = info.get("hess", info.get("hesse", info.get("hessian"))) if "inv_hessian" not in approx: approx["inv_hessian"] = info.get("inv_hess", info.get("inv_hesse", info.get("inv_hessian"))) if evaluator is not None: if "gradient" not in approx: approx["gradient"] = evaluator.last_gradient if "hessian" not in approx: approx["hessian"] = evaluator.last_hessian approx = Approximations(**approx) return approx def _input_convert_params(self, params): return ParamHolder((p, {"value": v}) for p, v in params.items()) def _check_get_uncached_params(self, params, method_name, cl): uncached = [] for p in params: errordict = self.params[p].get(method_name) # cl is < 1 and gets very close. The closer, the more it matters -> scale tolerance by it if errordict is not None and not math.isclose(errordict["cl"], cl, abs_tol=3e-3 * (1 - cl)): msg = ( f"Error with name {method_name} already exists in {self!r} with a different" f" convidence level of {errordict['cl']} instead of the requested {cl}." f" Use a different name." ) raise NameError( msg, ) uncached.append(p) return uncached def _create_minuit_instance(self): minuit = self._cache_minuit from zfit.minimizers.minimizer_minuit import Minuit if minuit is None: if isinstance(self.minimizer, Minuit): minuit = self.minimizer._minuit_minimizer else: minimizer = Minuit(tol=self.minimizer.tol, verbosity=0, name="ZFIT_TMP_UNCERTAINITIES") minuit, _, _ = minimizer._make_minuit(loss=self.loss, params=self.params, init=self) self._cache_minuit = minuit return minuit
[docs] @classmethod def from_ipopt( cls, loss: ZfitLoss, params: Iterable[ZfitParameter], problem: ipyopt.Problem, minimizer: zfit.minimize.IpyoptV1, valid: bool, values: np.ndarray, message: str | None, converged: bool | None, edm: zfit.minimizers.termination.CriterionNotAvailable | float, niter: int | None, fminopt: float | None, status: int | None, criterion: zfit.minimizers.termination.ConvergenceCriterion, evaluator: zfit.minimizers.evaluation.LossEval | None, ) -> FitResult: """Create a ``FitResult`` from an ipopt minimization. Args: loss: |@doc:result.init.loss| The loss function that was minimized. Usually, but not necessary, contains also the pdf, data and constraints. |@docend:result.init.loss| params: |@doc:result.init.params| Result of the fit where each :py:class:`~zfit.Parameter` key has the value from the minimum found by the minimizer. |@docend:result.init.params| problem: |@doc:result.init.problem||@docend:result.init.problem| minimizer: |@doc:result.init.minimizer| Minimizer that was used to obtain this ``FitResult`` and will be used to calculate certain errors. If the minimizer is state-based (like "iminuit"), then this is a copy and the state of other ``FitResults`` or of the *actual* minimizer that performed the minimization won't be altered. |@docend:result.init.minimizer| valid: |@doc:result.init.valid| Indicating whether the result is valid or not. This is the strongest indication and serves as the global flag. The reasons why a result may be invalid can be arbitrary, including but not exclusive: - parameter(s) at the limit - maxiter reached without proper convergence - the minimizer maybe even converged but it is known that this is only a local minimum To indicate the reason for the invalidity, pass a message. |@docend:result.init.valid| values: |@doc:result.init.values| Values of the parameters at the found minimum. |@docend:result.init.values| message: |@doc:result.init.message| Human-readable message to indicate the reason if the fitresult is not valid. If the fit is valid, the message (should) be an empty string (or None), otherwise, it should denote the reason for the invalidity. |@docend:result.init.message| converged: |@doc:result.init.converged| Whether the fit has successfully converged or not. The result itself can still be an invalid minimum such as if the parameters are at or close to the limits or in case another minimum is found. |@docend:result.init.converged| edm: |@doc:result.init.edm| The estimated distance to minimum which is the criterion value at the minimum. |@docend:result.init.edm| niter: |@doc:result.init.niter| Approximate number of iterations ~= number of function evaluations ~= number of gradient evaluations. This is an approximated value and the exact meaning can differ between different minimizers. |@docend:result.init.niter| fminopt: |@doc:result.init.fmin| Value of the function at the minimum. |@docend:result.init.fmin| status: |@doc:result.init.status| A status code (if available) that describes the minimization termination. 0 means a valid termination. |@docend:result.init.status| criterion: |@doc:result.init.criterion| Criterion that was used during the minimization. This determines the estimated distance to the minimum (edm) |@docend:result.init.criterion| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| Returns: ``zfit.minimize.FitResult``: """ info = {"problem": problem} params = dict(zip(params, values)) valid = valid if converged is None else valid and converged if evaluator is not None: valid = valid and not evaluator.maxiter_reached return cls( params=params, loss=loss, fminopt=fminopt, edm=edm, message=message, criterion=criterion, info=info, valid=valid, converged=converged, niter=niter, status=status, minimizer=minimizer, evaluator=evaluator, )
[docs] @classmethod def from_minuit( cls, loss: ZfitLoss, params: Iterable[ZfitParameter], minuit: iminuit.Minuit, minimizer: ZfitMinimizer | iminuit.Minuit, valid: bool | None, values: np.ndarray | None = None, message: str | None = None, converged: bool | None = None, edm: None | (zfit.minimizers.termination.CriterionNotAvailable | float) = None, niter: int | None = None, fminopt: float | None = None, status: int | None = None, criterion: zfit.minimizers.termination.ConvergenceCriterion | None = None, evaluator: zfit.minimizers.evaluation.LossEval | None = None, ) -> FitResult: """Create a `FitResult` from a :py:class:`~iminuit.util.MigradResult` returned by :py:meth:`iminuit.Minuit.migrad` and a iminuit :py:class:`~iminuit.Minuit` instance with the corresponding zfit objects. Args: loss: zfit Loss that was minimized. params: Iterable of the zfit parameters that were floating during the minimization. minuit: Return value of the iminuit migrad command, the instance of :class:`iminuit.Minuit` minimizer: Instance of the zfit Minuit minimizer that was used to minimize the loss. valid: |@doc:result.init.valid| Indicating whether the result is valid or not. This is the strongest indication and serves as the global flag. The reasons why a result may be invalid can be arbitrary, including but not exclusive: - parameter(s) at the limit - maxiter reached without proper convergence - the minimizer maybe even converged but it is known that this is only a local minimum To indicate the reason for the invalidity, pass a message. |@docend:result.init.valid| values: |@doc:result.init.values| Values of the parameters at the found minimum. |@docend:result.init.values| message: |@doc:result.init.message| Human-readable message to indicate the reason if the fitresult is not valid. If the fit is valid, the message (should) be an empty string (or None), otherwise, it should denote the reason for the invalidity. |@docend:result.init.message| converged: |@doc:result.init.converged| Whether the fit has successfully converged or not. The result itself can still be an invalid minimum such as if the parameters are at or close to the limits or in case another minimum is found. |@docend:result.init.converged| edm: |@doc:result.init.edm| The estimated distance to minimum which is the criterion value at the minimum. |@docend:result.init.edm| niter: |@doc:result.init.niter| Approximate number of iterations ~= number of function evaluations ~= number of gradient evaluations. This is an approximated value and the exact meaning can differ between different minimizers. |@docend:result.init.niter| fminopt: |@doc:result.init.fmin| Value of the function at the minimum. |@docend:result.init.fmin| status: |@doc:result.init.status| A status code (if available) that describes the minimization termination. 0 means a valid termination. |@docend:result.init.status| criterion: |@doc:result.init.criterion| Criterion that was used during the minimization. This determines the estimated distance to the minimum (edm) |@docend:result.init.criterion| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| Returns: ``zfit.minimize.FitResult``: A `FitResult` as if zfit Minuit was used. """ from .minimizer_minuit import Minuit from .termination import EDM if not isinstance(minimizer, Minuit): if isinstance(minimizer, iminuit.Minuit): minimizer_new = Minuit() minimizer_new._minuit_minimizer = minimizer minimizer = minimizer_new else: msg = f"Minimizer {minimizer} not supported. Use `Minuit` from zfit or from iminuit." raise ValueError(msg) params_result = list(minuit.params) fmin_object = minuit.fmin minuit_converged = not fmin_object.is_above_max_edm converged = minuit_converged if converged is None else (converged and minuit_converged) niter = fmin_object.nfcn if niter is None else niter info = { "n_eval": niter, # 'grad': result['jac'], # 'message': result['message'], "minuit": minuit, "original": fmin_object, } if fmin_object.has_covariance: info["inv_hessian"] = np.array(minuit.covariance) edm = fmin_object.edm if edm is None else edm if criterion is None: criterion = EDM(tol=minimizer.tol, loss=loss, params=params) criterion.last_value = edm fminopt = fmin_object.fval if fminopt is None else fminopt minuit_valid = fmin_object.is_valid valid = minuit_valid if valid is None else minuit_valid and valid if evaluator is not None: valid = valid and not evaluator.maxiter_reached if values is None: values = (res.value for res in params_result) params = dict(zip(params, values)) return cls( params=params, edm=edm, fminopt=fminopt, info=info, loss=loss, niter=niter, converged=converged, status=status, message=message, valid=valid, criterion=criterion, minimizer=minimizer, evaluator=evaluator, )
[docs] @classmethod def from_scipy( cls, loss: ZfitLoss, params: Iterable[ZfitParameter], result: scipy.optimize.OptimizeResult, minimizer: ZfitMinimizer, message: str | None, valid: bool, criterion: ConvergenceCriterion, edm: float | None = None, niter: int | None = None, evaluator: zfit.minimize.LossEval | None = None, ) -> FitResult: """Create a ``FitResult from a SciPy `~scipy.optimize.OptimizeResult`. Args: loss: |@doc:result.init.loss| The loss function that was minimized. Usually, but not necessary, contains also the pdf, data and constraints. |@docend:result.init.loss| params: |@doc:result.init.params| Result of the fit where each :py:class:`~zfit.Parameter` key has the value from the minimum found by the minimizer. |@docend:result.init.params| result: Result of the SciPy optimization. minimizer: |@doc:result.init.minimizer| Minimizer that was used to obtain this ``FitResult`` and will be used to calculate certain errors. If the minimizer is state-based (like "iminuit"), then this is a copy and the state of other ``FitResults`` or of the *actual* minimizer that performed the minimization won't be altered. |@docend:result.init.minimizer| message: |@doc:result.init.message| Human-readable message to indicate the reason if the fitresult is not valid. If the fit is valid, the message (should) be an empty string (or None), otherwise, it should denote the reason for the invalidity. |@docend:result.init.message| edm: |@doc:result.init.edm| The estimated distance to minimum which is the criterion value at the minimum. |@docend:result.init.edm| niter: |@doc:result.init.niter| Approximate number of iterations ~= number of function evaluations ~= number of gradient evaluations. This is an approximated value and the exact meaning can differ between different minimizers. |@docend:result.init.niter| valid: |@doc:result.init.valid| Indicating whether the result is valid or not. This is the strongest indication and serves as the global flag. The reasons why a result may be invalid can be arbitrary, including but not exclusive: - parameter(s) at the limit - maxiter reached without proper convergence - the minimizer maybe even converged but it is known that this is only a local minimum To indicate the reason for the invalidity, pass a message. |@docend:result.init.valid| criterion: |@doc:result.init.criterion| Criterion that was used during the minimization. This determines the estimated distance to the minimum (edm) |@docend:result.init.criterion| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| Returns: `zfit.minimize.FitResult`: """ result_values = result["x"] if niter is None: niter = result.get("nit") converged = result.get("success", valid) status = result["status"] if message is None and (not converged or not valid): message = result.get("message") grad = result.get("grad") info = { "n_eval": result["nfev"], "n_iter": niter, "niter": niter, "grad": result.get("jac") if grad is None else grad, "message": message, "evaluator": evaluator, "original": result, } approx = { "params": params, "gradient": info.get("grad"), } if info.get("niter", 0) > 25: # unreliable if too few iterations, fails for EDM inv_hesse = result.get("hess_inv") if isinstance(inv_hesse, LbfgsInvHessProduct): inv_hesse = inv_hesse.todense() hesse = info.get("hesse") info["inv_hesse"] = inv_hesse info["hesse"] = hesse approx["hessian"] = hesse approx["inv_hessian"] = inv_hesse fminopt = result["fun"] params = dict(zip(params, result_values)) if evaluator is not None: valid = valid and not evaluator.maxiter_reached return cls( params=params, edm=edm, fminopt=fminopt, info=info, approx=approx, converged=converged, status=status, message=message, valid=valid, niter=niter, loss=loss, minimizer=minimizer, criterion=criterion, evaluator=evaluator, )
[docs] @classmethod def from_nlopt( cls, loss: ZfitLoss, opt, params: Iterable[ZfitParameter], minimizer: ZfitMinimizer | iminuit.Minuit, valid: bool | None, values: np.ndarray | None = None, message: str | None = None, converged: bool | None = None, edm: None | (zfit.minimizers.termination.CriterionNotAvailable | float) = None, niter: int | None = None, fminopt: float | None = None, status: int | None = None, criterion: zfit.minimizers.termination.ConvergenceCriterion | None = None, evaluator: zfit.minimizers.evaluation.LossEval | None = None, inv_hessian: np.ndarray | None = None, hessian: np.ndarray | None = None, ) -> FitResult: """Create a ``FitResult`` from an NLopt optimizer. Args: loss: |@doc:result.init.loss| The loss function that was minimized. Usually, but not necessary, contains also the pdf, data and constraints. |@docend:result.init.loss| opt: Optimizer instance of NLopt params: |@doc:result.init.params| Result of the fit where each :py:class:`~zfit.Parameter` key has the value from the minimum found by the minimizer. |@docend:result.init.params| minimizer: |@doc:result.init.minimizer| Minimizer that was used to obtain this ``FitResult`` and will be used to calculate certain errors. If the minimizer is state-based (like "iminuit"), then this is a copy and the state of other ``FitResults`` or of the *actual* minimizer that performed the minimization won't be altered. |@docend:result.init.minimizer| valid: |@doc:result.init.valid| Indicating whether the result is valid or not. This is the strongest indication and serves as the global flag. The reasons why a result may be invalid can be arbitrary, including but not exclusive: - parameter(s) at the limit - maxiter reached without proper convergence - the minimizer maybe even converged but it is known that this is only a local minimum To indicate the reason for the invalidity, pass a message. |@docend:result.init.valid| values: |@doc:result.init.values| Values of the parameters at the found minimum. |@docend:result.init.values| message: |@doc:result.init.message| Human-readable message to indicate the reason if the fitresult is not valid. If the fit is valid, the message (should) be an empty string (or None), otherwise, it should denote the reason for the invalidity. |@docend:result.init.message| converged: |@doc:result.init.converged| Whether the fit has successfully converged or not. The result itself can still be an invalid minimum such as if the parameters are at or close to the limits or in case another minimum is found. |@docend:result.init.converged| edm: |@doc:result.init.edm| The estimated distance to minimum which is the criterion value at the minimum. |@docend:result.init.edm| niter: |@doc:result.init.niter| Approximate number of iterations ~= number of function evaluations ~= number of gradient evaluations. This is an approximated value and the exact meaning can differ between different minimizers. |@docend:result.init.niter| fminopt: |@doc:result.init.fmin| Value of the function at the minimum. |@docend:result.init.fmin| status: |@doc:result.init.status| A status code (if available) that describes the minimization termination. 0 means a valid termination. |@docend:result.init.status| criterion: |@doc:result.init.criterion| Criterion that was used during the minimization. This determines the estimated distance to the minimum (edm) |@docend:result.init.criterion| evaluator: |@doc:result.init.evaluator| Loss evaluator that was used during the minimization and that may contain information about the last evaluations of the gradient etc. which can serve as approximations. |@docend:result.init.evaluator| inv_hessian: The (approximated) inverse hessian matrix. hessian: The (approximated) hessian matrix. Returns: zfit.minimizers.fitresult.FitResult: """ converged = converged if converged is None else bool(converged) param_dict = dict(zip(params, values)) if fminopt is None: fminopt = opt.last_optimum_value() status_nlopt = opt.last_optimize_result() if status is None: status = status_nlopt niter = opt.get_numevals() if niter is None else niter converged = 1 <= status_nlopt <= 4 and converged is not False messages = { 1: "NLOPT_SUCCESS", 2: "NLOPT_STOPVAL_REACHED", 3: "NLOPT_FTOL_REACHED", 4: "NLOPT_XTOL_REACHED", 5: "NLOPT_MAXEVAL_REACHED", 6: "NLOPT_MAXTIME_REACHED", -1: "NLOPT_FAILURE", -2: "NLOPT_INVALID_ARGS", -3: "NLOPT_OUT_OF_MEMORY", -4: "NLOPT_ROUNDOFF_LIMITED", -5: "NLOPT_FORCED_STOP", } message_nlopt = messages[status_nlopt] info = { "n_eval": niter, "niter": niter, "message": message_nlopt, "original": status, "evaluator": evaluator, "status": status, } if message is None: message = message_nlopt valid = valid and converged if evaluator is not None: valid = valid and not evaluator.maxiter_reached approx = {} if inv_hessian is None and hessian is None and evaluator is not None: hessian = evaluator.last_hessian # if hessian is not None: # TODO: remove? # inv_hessian = np.linalg.inv(hessian) if inv_hessian is not None: info["inv_hesse"] = inv_hessian approx["inv_hessian"] = inv_hessian return cls( params=param_dict, edm=edm, fminopt=fminopt, status=status, converged=converged, info=info, niter=niter, valid=valid, loss=loss, minimizer=minimizer, criterion=criterion, message=message, evaluator=evaluator, )
@property def approx(self) -> Approximations: return self._approx @property def params( self, ) -> Mapping[ZfitIndependentParameter, Mapping[str, Mapping[str, object]]]: return self._params @property def values(self) -> Mapping[str | ZfitParameter, float]: return self._values @property def criterion(self) -> ConvergenceCriterion: return self._criterion @property def message(self) -> str: return self._message @property def edm(self) -> float: """The estimated distance to the minimum. Returns: Numeric """ return self._edm @property def minimizer(self) -> ZfitMinimizer: return self._minimizer @property def loss(self) -> ZfitLoss: # TODO(Mayou36): this is currently a reference, should be a copy of the loss? return self._loss @property def fminopt(self) -> float: """Function value with possible optimizations at the minimum, namely constant subtraction. Returns: Numeric """ return self._fminopt @property def fmin(self) -> float: """Function value, fully evaluated, at the minimum. Returns: Numeric """ return self._fmin @property @deprecated( None, "Use `fmin` instead which now returns the full minimum value. This will be removed in the future.", ) def fminfull(self) -> float: """Function value, fully evaluated, at the minimum. Returns: Numeric """ return self.fmin @property def status(self): return self._status @property def info(self) -> Mapping[str, object]: return self._info @property def converged(self) -> bool: return self._converged @property def valid(self) -> bool: return self._valid and not self.params_at_limit and self.converged @property def params_at_limit(self) -> bool: return self._params_at_limit @contextlib.contextmanager def _input_check_reset_params(self, params): params = self._input_check_params(params=params) old_values = np.asarray(params) try: yield params except Exception: warnings.warn( "Exception occurred, parameter values are not reset and in an arbitrary, last" " used state. If this happens during normal operation, make sure you reset the values.", RuntimeWarning, stacklevel=3, ) raise set_values(params=params, values=old_values, allow_partial=True) # TODO: or set? def _input_check_params(self, params): return convert_to_container(params) if params is not None else list(self.params.keys())
[docs] @deprecated_args(None, "Use `name` instead", "error_name") def hesse( self, params: ParamsTypeOpt = None, method: str | Callable | None = None, cl: float | None = None, name: str | bool | None = None, # DEPRECATED error_name: str | None = None, ) -> dict[ZfitIndependentParameter, dict]: r"""Calculate for `params` the symmetric error using the Hessian/covariance matrix. This method estimates the covariance matrix using the inverse of the Hessian matrix. The assumption is that the loss profile - usually a likelihood or a :math:`\chi^2` - is hyperbolic. This is usually the case for fits with many observations, i.e. it is exact in the asymptotic limit. If the loss profile is not hyperbolic, another method, "zfit_error" or "minuit_minos" should be used. **Weights** Weighted likelihoods are a special class of likelihoods as they are not an actual likelihood. However, the minimum is still valid, however the profile is not a proper likelihood. Therefore, corrections will be automatically applied to the Hessian uncertainty estimation in order to correct for the effects in the weights. The corrections used are "asymptotically correct" and are described in `Parameter uncertainties in weighted unbinned maximum likelihood fits`<https://doi.org/10.1140/epjc/s10052-022-10254-8>` by Christoph Langenbruch. Since this method uses the jacobian matrix, it takes significantly longer to calculate than witout weights. Args: params: The parameters to calculate the Hessian symmetric error. If None, use all parameters. method: the method to calculate the covariance matrix. Can be {'minuit_hesse', 'hesse_np', 'approx'} or a callable. cl: Confidence level for the error. If None, use the default value of 0.68. name: The name for the error in the dictionary. This will be added to the information collected in params under ``params[p][name]`` where p is a Parameter. If the name is `False`, it won't be added and only returned. Defaulst to `'hesse'`. Returns: Result of the hessian (symmetric) error as dict with each parameter holding the error dict {'error': sym_error}. So given param_a (from zfit.Parameter(.)) `error_a = result.hesse(params=param_a)[param_a]['error']` error_a is the hessian error. """ # for compatibility with `errors` cl = 0.68268949 if cl is None else cl # scipy.stats.chi2(1).cdf(1) if cl >= 1: msg = f"cl is the confidence limit and has to be < 1, not {cl}" raise ValueError(msg) if method is None: # LEGACY START method = self._default_hesse from zfit.minimizers.minimizer_minuit import Minuit if isinstance(self.minimizer, Minuit): method = "minuit_hesse" # LEGACY END # Deprecated name if error_name is not None: name = error_name name_warning_triggered = False if name is None: if not isinstance(method, str): msg = "Need to specify `name` or use a string as `method`" raise ValueError(msg) name = "hesse" with self._input_check_reset_params(params) as params: uncached_params = self._check_get_uncached_params(params=params, method_name=name, cl=cl) if uncached_params: error_dict = self._hesse(params=uncached_params, method=method, cl=cl) if any(val["error"] is None for val in error_dict.values()): return {} for p in error_dict: error_dict[p]["cl"] = cl if name: self._cache_errors(name=name, errors=error_dict) else: error_dict = {} error_dict.update({p: self.params[p][name] for p in params if p not in uncached_params}) if name_warning_triggered: error_dict.update({p: self.params[p][method] for p in params if p not in uncached_params}) return {p: error_dict[p] for p in params}
def _cache_errors(self, name, errors): for param, error in errors.items(): self.params[param][name] = error def _hesse(self, params, method, cl): pseudo_sigma = scipy.stats.chi2(1).ppf(cl) ** 0.5 covariance_dict = self.covariance(params, method, as_dict=True) return { p: { "error": ( float(covariance_dict[(p, p)]) ** 0.5 * pseudo_sigma if covariance_dict[(p, p)] is not None else None ) } for p in params }
[docs] def error( self, params: ParamsTypeOpt = None, # noqa: ARG002 method: str | Callable | None = None, # noqa: ARG002 error_name: str | None = None, # noqa: ARG002 sigma: float = 1.0, # noqa: ARG002 ) -> dict: r""".. deprecated:: unknown Use :func:`errors` instead. Args: params: The parameters or their names to calculate the errors. If `params` is `None`, use all *floating* parameters. method: The method to use to calculate the errors. Valid choices are {'minuit_minos'} or a Callable. sigma: Errors are calculated with respect to `sigma` std deviations. The definition of 1 sigma depends on the loss function and is defined there. For example, the negative log-likelihood (without the factor of 2) has a correspondents of :math:`\Delta` NLL of 1 corresponds to 1 std deviation. error_name: The name for the error in the dictionary. Returns: A `dict` containing as keys the parameter names and as value a `dict` which contains (next to probably more things) two keys 'lower' and 'upper', holding the calculated errors. Example: result['par1']['upper'] -> the asymmetric upper error of 'par1' """ msg = "Use `errors` instead of `error`." raise BreakingAPIChangeError(msg)
[docs] @deprecated_args(None, "Use name instead.", "error_name") def errors( self, params: ParamsTypeOpt = None, method: str | Callable | None = None, name: str | None = None, cl: float | None = None, *, sigma=None, error_name: str | None = None, ) -> tuple[dict, None | FitResult]: r"""Calculate and set for `params` the asymmetric error using the set error method. Args: params: The parameters or their names to calculate the errors. If `params` is `None`, use all *floating* parameters. method: The method to use to calculate the errors. Valid choices are {'minuit_minos', 'zfit_errors'} or a Callable. cl: Uncertainties are calculated with respect to the confidence level cl. The default is 68.3%. For example, the negative log-likelihood (without the factor of 2) has a correspondents of :math:`\Delta` NLL of 1 corresponds to 1 std deviation. sigma: Errors are calculated with respect to `sigma` std deviations. The definition of 1 sigma depends on the loss function and is defined there. name: The name for the error in the dictionary. Defaults to `errors` Returns: A `dict` containing as keys the parameter and as value a `dict` which contains (next to often more things) two keys 'lower' and 'upper', holding the calculated errors. Furthermore, it has `cl` to indicate the convidence level the uncertainty was calculated with. Example: result[par1]['upper'] -> the asymmetric upper error of 'par1' """ # Deprecated name if error_name is not None: name = error_name if sigma is not None: if cl is not None: msg = "Cannot define sigma and cl, use only one." raise ValueError(msg) cl = scipy.stats.chi2(1).cdf(sigma) if cl is None: cl = 0.68268949 # scipy.stats.chi2(1).cdf(1) if method is None: method = self._default_error if name is None: if not isinstance(method, str): msg = "Need to specify `name` or use a string as `method`" raise ValueError(msg) name = "errors" if method == "zfit_error": warnings.warn( "'zfit_error' is still somewhat new. If it fails, please report it here:" " https://github.com/zfit/zfit/issues/new?assignees=&labels=bug&template" "=bug_report.md&title=zfit%20error%20fails.", ExperimentalFeatureWarning, stacklevel=2, ) params = self._input_check_params(params) with self._input_check_reset_params(self.params.keys()): uncached_params = self._check_get_uncached_params(params=params, method_name=name, cl=cl) new_result = None if uncached_params: error_dict, new_result = self._error(params=uncached_params, method=method, cl=cl) for p in error_dict: error_dict[p]["cl"] = cl self._cache_errors(name=name, errors=error_dict) if new_result is not None: msg = "Invalid, a new minimum was found." self._cache_errors(name=name, errors={p: msg for p in params}) self._valid = False self._message = msg new_result._cache_errors(name=name, errors=error_dict) all_errors = {p: self.params[p][name] for p in params} return all_errors, new_result
def _error(self, params, method, cl): if not callable(method): try: method = self._error_methods[method] except KeyError: msg = f"The following method is not a valid, implemented method: {method}. Use one of {self._error_methods.keys()}" raise KeyError(msg) from None return method(result=self, params=params, cl=cl)
[docs] def covariance( self, params: ParamsTypeOpt = None, method: str | Callable | None = None, as_dict: bool = False, ): """Calculate the covariance matrix for `params`. Args: params: The parameters to calculate the covariance matrix. If `params` is `None`, use all *floating* parameters. method: The method to use to calculate the covariance matrix. Valid choices are {'minuit_hesse', 'hesse_np'} or a Callable. as_dict: Default `False`. If `True` then returns a dictionnary. Returns: 2D `numpy.array` of shape (N, N); `dict`(param1, param2) -> covariance if `as_dict == True`. """ if method is None: method = self._default_hesse if method not in self._covariance_dict: with self._input_check_reset_params(params) as params: self._covariance_dict[method] = self._covariance(method=method) params = self._input_check_params(params) covariance = {k: self._covariance_dict[method].get(k) for k in itertools.product(params, params)} if as_dict: return covariance return dict_to_matrix(params, covariance)
def _covariance(self, method): if not callable(method): if method not in self._hesse_methods: msg = f"The following method is not a valid, implemented method: {method}. Use one of {self._hesse_methods.keys()}" raise KeyError(msg) method = self._hesse_methods[method] params = list(self.params.keys()) if any(isinstance(data, ZfitData) and data.weights is not None for data in self.loss.data): return covariance_with_weights(method=method, result=self, params=params) return method(result=self, params=params)
[docs] def correlation( self, params: ParamsTypeOpt = None, method: str | Callable | None = None, as_dict: bool = False, ): """Calculate the correlation matrix for `params`. Args: params: The parameters to calculate the correlation matrix. If `params` is `None`, use all *floating* parameters. method: The method to use to calculate the correlation matrix. Valid choices are {'minuit_hesse', 'hesse_np'} or a Callable. as_dict: Default `False`. If `True` then returns a dictionnary. Returns: 2D `numpy.array` of shape (N, N); `dict`(param1, param2) -> correlation if `as_dict == True`. """ covariance = self.covariance(params=params, method=method, as_dict=False) correlation = covariance_to_correlation(covariance) if as_dict: params = self._input_check_params(params) return matrix_to_dict(params, correlation) return correlation
[docs] def freeze(self): """Freeze the result to make it pickleable and convert all TensorFlow elements to names (parameters) or arrays. After this, no more uncertainties or covariances can be calculated. The already calculated ones remain however. Parameters can be accessed by their string name. """ self._loss = self.loss.name self._minimizer = self.minimizer.name self._criterion = self.criterion.name self._evaluator = None self.approx.freeze() self._covariance_dict = { k: {(p[0].name, p[1].name): v for p, v in d.items()} for k, d in self._covariance_dict.items() } self._values = ValuesHolder({p.name: self.values[p] for p in self.params}) self._params = ParamHolder({k.name: v for k, v in self.params.items()}) if "minuit" in self.info: self.info["minuit"] = "Minuit_frozen" if "problem" in self.info: try: import ipyopt except ImportError: pass else: if isinstance(self.info["problem"], ipyopt.Problem): self.info["problem"] = "ipyopt_frozen" if "evaluator" in self.info: self.info["evaluator"] = "evaluator_frozen" self._cache_minuit = None
def __str__(self): string = Style.BRIGHT + "FitResult" + Style.NORMAL + f" of\n{self.loss} \nwith\n{self.minimizer}\n\n" string += tabulate( [ [ color_on_bool(self.valid), color_on_bool(self.converged, on_true=False), color_on_bool(self.params_at_limit, on_true=colored.bg(9), on_false=False), format_value(self.edm, highprec=False), f" {self._fmin:.2f} | {format_value(self.fminopt)}", ] ], [ "valid", "converged", "param at limit", "edm", "approx. fmin (full | opt.)", ], tablefmt="fancy_grid", disable_numparse=True, colalign=["center", "center", "center", "center", "right"], ) string += "\n\n" + Style.BRIGHT + "Parameters\n" + Style.NORMAL string += str(self.params) return string def _repr_pretty_(self, p, cycle): if cycle: p.text(self.__repr__()) return p.text(self.__str__()) def __enter__(self): self._tmp_old_param_values.append(znp.asarray(tuple(self.params.keys()))) self.update_params() return self def __exit__(self, exc_type, exc_val, exc_tb): old_vals = self._tmp_old_param_values.pop() set_values(tuple(self.params.keys()), old_vals)
[docs] def update_params(self) -> typing.Self: """Update the parameters of the result to the current values. Usually to be used chained with a ``minimize`` call to update the parameters to the current values. (currently, the parameters are updated to the values of the result, this might change in the future. To enable this expected behavior, use ``zfit.run.experimental_disable_param_update()``. .. code-block:: python result = minimizer.minimize(loss).update_params() # update the parameters to the current values """ set_values(self) return self
def covariance_to_correlation(covariance): diag = np.diag(1 / np.diag(covariance) ** 0.5) return np.matmul(diag, np.matmul(covariance, diag)) def format_value(value, highprec=True): m_error_class = iminuit.util.MError # if iminuit is not available (maybe in the future?), use dict instead if isinstance(value, dict) and "error" in value: value = value["error"] value = f"{value:> 6.2g}" value = f'+/-{" " * (8 - len(value))}' + value if isinstance(value, m_error_class) or (isinstance(value, dict) and "lower" in value and "upper" in value): if isinstance(value, m_error_class): lower = value.lower upper = value.upper else: lower = value["lower"] upper = value["upper"] lower_sign = f"{np.sign(lower): >+}"[0] upper_sign = f"{np.sign(upper): >+}"[0] lower, upper = f"{np.abs(lower): >6.2g}", f"{upper: >6.2g}" lower = lower_sign + " " * (7 - len(lower)) + lower upper = upper_sign + " " * (7 - len(upper)) + upper # lower += " t" * (11 - len(lower)) value = lower + " " * 3 + upper if isinstance(value, float): value = f"{value:> 6.7g}" if highprec else f"{value:> 6.2g}" return value def color_on_bool(value, on_true=None, on_false=None): if on_true is None: on_true = colored.bg(10) if on_false is None: on_false = colored.bg(9) if not value and on_false: value_add = on_false elif value and on_true: value_add = on_true else: value_add = "" return value_add + str(value) + Style.RESET_ALL class ListWithKeys(collections.UserList): __slots__ = ("_initdict",) def __init__(self, initdict) -> None: super().__init__(initlist=initdict.values()) self._initdict = initdict def __getitem__(self, item): if isinstance(item, ZfitParameter): return self._initdict[item] return super().__getitem__(item) def keys(self): return self._initdict.keys() def values(self): return self._initdict.values() def items(self): return self._initdict.items() class ValuesHolder(NameToParamGetitem, ListWithKeys): __slots__ = () class ParamHolder(NameToParamGetitem, collections.UserDict): def __str__(self) -> str: order_keys = ["value", "hesse"] keys = OrderedSet() for pdict in self.values(): keys.update(OrderedSet(pdict)) order_keys = OrderedSet([key for key in order_keys if key in keys]) order_keys.update(keys) rows = [] for param, pdict in self.items(): name = param.name if isinstance(param, ZfitParameter) else param row = [name] row.extend(format_value(pdict.get(key, " ")) for key in order_keys) if isinstance(param, ZfitParameter): row.append( color_on_bool( bool(param.at_limit), on_true=colored.bg("light_red"), on_false=False, ) ) rows.append(row) order_keys = ["name", *list(order_keys), "at limit"] order_keys[order_keys.index("value")] = "value (rounded)" return tabulate(rows, order_keys, numalign="right", stralign="right", colalign=("left",))