Likelihood based fitting
"Fitting" simply means estimating some parameters of a model (or really a set of models) based on data. Likelihood-based fitting does this through the likelihood function.
In frequentist frameworks, this typically means doing maximum likelihood estimation. In bayesian frameworks, usually posterior distributions of the parameters are calculated from the likelihood.
Fitting Frameworks
Likelihood fits typically either follow a frequentist framework of maximum likelihood estimation, or a bayesian framework of updating estimates to find posterior distributions given the data.
Maximum Likelihood fits
A maximum likelihood fit means finding the values of the model parameters
These values provide point estimates for the parameter values.
Because the likelihood is equal to the probability of observing the data given the model, the maximum likelihood estimate finds the parameter values for which the data is most probable.
Bayesian Posterior Calculation
In a bayesian framework, the likelihood represents the probability of observing the data given the model and some prior probability distribution over the model parameters.
The prior probability of the parameters,
The posterior distribution
Methods for considering subsets of models
Often, one is interested in some particular aspect of a model. This may be for example information related to the parameters of interest, but not the nuisance parameters. In this case, one needs a method for specifying precisely what is meant by a model considering only those parameters of interest.
There are several methods for considering sub models which each have their own interpretations and use cases.
Conditioning
Conditional Sub-models can be made by simply restricting the values of some parameters.
The conditional likelihood of the parameters
Profiling
The profiled likelihood
In some sense, the profiled likelihood is the best estimate of the likelihood at every point
Marginalization
Marginalization is a procedure for producing a probability distribution
The marginalized probability
Marginalized likelihoods can also be defined, by their relationship to the probability distributions.
Parameter Uncertainties
Parameter uncertainties describe regions of parameter values which are considered reasonable parameter values, rather than single estimates. These can be defined either in terms of frequentist confidence regions or bayesian credibility regions.
In both cases the region is defined by a confidence or credibility level
The confidence or credibility regions are described by a set of points
Typically indicated as:
or, if symmetric intervals are used:
Frequentist Confidence Regions
Frequentist confidence regions are random variables of the observed data. These are very often the construction used to define the uncertainties reported on a parameter.
If the same experiment is repeated multiple times, different data will be osbserved each time and a different confidence set
From first principles, the intervals can be constructed using the Neyman construction.
In practice, the likelihood can be used to construct confidence regions for a set of parameters
i.e. the ratio of the profile likelihood at point
Each point
The cutoff value
Under some conditions, the value of
Constructing Frequentist Confidence Regions in Practice
When a single fit is performed by some numerical minimization program and parameter values are reported along with some uncertainty values, they are usually reported as frequentist intervals. The MINUIT minimizer which evaluates likelihood functions has two methods for estimating parameter uncertainties.
These two methods are the most commonly used methods for estimating confidence regions in a fit; they are the minos method, and the hessian method.
In both cases, Wilk's theorem is assumed to hold at all points in parameter space, such that
When
The Minos method for estimating confidence regions
In the minos method, once the best fit point
Following this procedure,
The search is performed in both directions, away from the best fit value of the parameter and the two crossings are taken as the borders of the confidence region.
This procedure has to be followed sepately for each parameter
The Hessian method for estimating confidence regions
The Hessian method relies on the second derivatives (i.e. the hessian) of the likelihood at the best fit point.
By assuming that the shape of the likelihood function is well described by its second-order approximation, the values at which
By computing and then inverting the full hessian matrix, all individual confidence regions and the full covariance matrix are determined. By construction, this method always reports symmetric confidence intervals, as it assumes that the likelihood is well described by a second order expansion.
Bayesian Credibility Regions
Often the full posterior probability distribution is summarized in terms of some credible region which contains some specified portion of the posterior probability of the parameter.
The credible region represents a region in which the bayesian probability of the parameter being in that region is equal to the chosen Credibility Level.