#Principles and pathology of orthodox statistics.
Given a fixed prior and model, a Bayesean estimator (MAP, posterior mean, or minimizing some other posterior risk insetad of MSE, or of any type at all) is still an estimator - it is a number
"This decreases monotonically from
Short summary of derivation of Cramer-Rao bound 17.30: apply Cauchy-Schwarts to the "covarince dot-product" of the score and the estimator. Slightly longer summary is as follows.
The formula 17.29 just expresses the following: covariance is a dot product (positive definite bilinear form) on the space of zero-mean random variables with finite second moments; the "(Cauchy-Buniakowski-)Schwartz" inequality for this dot product aka
The cool thing is that the score -- the derivative (in the parameter) of the log-likeliehood is -- a zero mean random variable, whose variance is the Fisher information, and whose covariance with an estimator
The equality condition is that the estimator and the score are proportional: that is, that the parameter is the parameter of exponential family, aka the "maximal entropy" family.
Here is how Jaynes's criticism of "rejection of null hypothesis" statistical method transfers to a criticism of mathematical technique of "proof by contradiction" (quote from 15.5.1 with italics for modifications)""
This is the kind of logic that underlies all proofs by contradiction. In order to argue for a conclusion C, one does it indirectly: invent a hypothesis "not C" that denies C, then argue against "not C" in a way that makes no reference to C at all (that is, using only deductions based on "not C"). To see how far this procedure takes us from elementary logic, suppose we decide that the conclusion holds; that is, we reject "not C". Surely, we must also reject reasoning based on "not C"; but then what was the logical justification for the deduction? Proof by contradiction saws off its own limb.