All Questions
Tagged with sufficient-statistics probability
29
questions
1
vote
1
answer
106
views
Formal definition of sufficient statistic
Let $(\Omega_X,\mathcal{F}_X)$ and $(\Omega _T,\mathcal{F}_T)$ be measurable spaces. Let $\mathfrak{M}$ be a family of probability measures on $(\Omega_X,\mathcal{F}_X)$. Let $X:\Omega\to \Omega _X$ ...
1
vote
0
answers
69
views
Likelihood ratio as minimal sufficient statistics in infinite parameter space
I just read a question from here (Likelihood ratio minimal sufficient) and have some thoughts. Let me restate the question first:
Consider a family of density functions $f(x|\theta)$ where the ...
1
vote
0
answers
37
views
Sufficient Statistic for a finite family of Normal distributions
Suppose we have a finite family of normal distributions $P=\{N(0, 1), N(0, 2), N(1, 2), N(2, 2)\}$ and we want to find a sufficient statistic for this family. Intuitively it is clear that as the means ...
1
vote
1
answer
3k
views
Sufficient statistics for bernoulli distribution
Let $Y_1, \ldots, Y_n $ be a random sample of size $n$ where each $Y_i \sim \textrm{Bernoulli}(p), $ and
let $Y = \sum Y_i $ for $i = 1, \ldots, n.$
The estimator is $W= (Y+1)/(n+2). $
Is the ...
11
votes
1
answer
643
views
Is there a standard measure of the sufficiency of a statistic?
Given a parametrical model $f_\theta$ and a random sample $X = (X_1, \cdots, X_n)$ from this model,
a statistic $T(X)$ is sufficient if the distribution of $X$ given $T(X)$ doesn't depend on $\theta$.
...
2
votes
2
answers
365
views
Sufficient statistic $\sum_{j=1}^{n} |x_{j}|$ for laplace distribution
Let be $X_{1},\ldots , X_{n}$ random variables independent and identically distributed with density function:
$$ f_{\theta}(x)=\dfrac{1}{2}e^{-|x-\mu|}, \quad x,\mu \in \mathbb{R} $$
Find the joint ...
0
votes
0
answers
139
views
Subscripts for Expectations and variances in for estimators [duplicate]
Is there any significance for subscripts to E and Var?
For example, the risk function of an estimator $\delta(\mathbf x)$ of $\theta$ in my book is:
$$
R(\theta,\delta)=E_\theta[L(\theta,\delta(\...
1
vote
1
answer
78
views
Did I correctly apply the factorisation theorem in this example?
Suppose that we have a density $f(x,\theta)=c(\theta)\psi(x)\unicode{x1D7D9}(x \in]\theta,\theta+1[)$ and the random variable $\mathbf{X}=(X_1,\ldots,X_n)$ are independently identically distributed ...
3
votes
1
answer
1k
views
Understanding the Rao-Blackwell Theorem
I've been reading up a lot on the practical applications of the Rao-Blackwell theorem. I do understand how the Bias and Variance and MSE aspects of the theorem fall in place (i.e. the mathematical ...
3
votes
2
answers
566
views
Do sufficient statistics for parameters of interest depend on whether nuisance parameters are known?
The definition of sufficient statistic is as follows:
A statistic $T(X_1,...,X_n)$ is sufficient for parameter $\theta$ if the conditional distribution of $X_1,...,X_n$, given that $T=t$, does not ...
1
vote
1
answer
969
views
Sufficient statistic for the distribution of a random sample of Poisson distribution
Let $X_1,...,X_n$ be a random sample from a Poisson distribution with mean $\lambda$ and $T = \sum_{i=1}^n X_i $ . Show that the distribution of $X_1,...,X_n$ given T is independant of $\lambda$ so ...
1
vote
1
answer
3k
views
Sufficient estimator for Bernoulli distribution using the likelihood function theorem for sufficiency
Let $(X_1,X_2)$ be a random sample of two iid random variables, $X_1\sim Ber(\theta),\theta\in (0,1)$.
Use the following theorem to show that $\hat{\theta}=X_1+2X_2$ is sufficient.
Likelihood theorem ...
1
vote
1
answer
1k
views
Sufficient Statistic for Normal Distribution | Mean, Variance & Kurtosis
I have seen multiple times that a normal distribution is fully specified by mean and variance. It is obvious that the third moment is not necessary for a perfect normal distribution as it is 0. I ...
4
votes
1
answer
83
views
When family of DF's $\mathcal{P}$ fail to be dominated by a measure $\mu$
On the topic of minimal sufficient statistics, there is an important theorem which requires the family of probability distributions $\mathcal{P}$ is dominated by some measure $\mu$.
As I understand it,...
14
votes
1
answer
1k
views
Intuitive understanding of the Halmos-Savage theorem
The Halmos-Savage theorem says that for a dominated statistical model $(\Omega, \mathscr A, \mathscr P)$ a statistic $T: (\Omega, \mathscr A, \mathscr P)\to(\Omega', \mathscr A')$ is sufficient if (...