Publication | Open Access
Why Least Squares and Maximum Entropy? An Axiomatic Approach to Inference for Linear Inverse Problems
805
Citations
20
References
1991
Year
Mathematical ProgrammingEngineeringComputational ComplexityUncertainty QuantificationRegularization (Mathematics)Estimation TheoryKolmogorov ComplexityLinear Inverse ProblemsInformation TheoryComputational Learning TheoryInverse ProblemsComputer ScienceMaximum EntropyStatistical Learning TheoryAlgorithmic Information TheoryLeast SquaresComposition ConsistencyEntropyNatural AxiomsStatistical Inference
An attempt is made to determine the logically consistent rules for selecting a vector from any feasible set defined by linear constraints, when either all $n$-vectors or those with positive components or the probability vectors are permissible. Some basic postulates are satisfied if and only if the selection rule is to minimize a certain function which, if a "prior guess" is available, is a measure of distance from the prior guess. Two further natural postulates restrict the permissible distances to the author's $f$-divergences and Bregman's divergences, respectively. As corollaries, axiomatic characterizations of the methods of least squares and minimum discrimination information are arrived at. Alternatively, the latter are also characterized by a postulate of composition consistency. As a special case, a derivation of the method of maximum entropy from a small set of natural axioms is obtained.
| Year | Citations | |
|---|---|---|
Page 1
Page 1