Skip to main content

All Questions

1 vote
0 answers
42 views

Show that $T=\sum_{i=1}^n X_i$ is a sufficient statistic for $p$ [duplicate]

I try to use the definition of sufficient statistic to prove that Suppose that $X_1,\dots, X_n$ is an iid random sample from $X\sim \mathrm{Bernoulli}(p)$. Show that $T=\sum_{i=1}^n X_i$ is a ...
Hermi's user avatar
  • 747
0 votes
0 answers
32 views

Extending Minimal sufficient statistics to arbitrary dimension

I am wondering if the following reasoning is correct regarding minimal sufficiency and dimension. Given $X_1,\dots,X_n$ i.i.d. $N(\mu,1)$, we know that the sample mean $S = \bar{X}$ is a minimal ...
WeakLearner's user avatar
  • 1,501
1 vote
1 answer
3k views

Sufficient statistics for bernoulli distribution

Let $Y_1, \ldots, Y_n $ be a random sample of size $n$ where each $Y_i \sim \textrm{Bernoulli}(p), $ and let $Y = \sum Y_i $ for $i = 1, \ldots, n.$ The estimator is $W= (Y+1)/(n+2). $ Is the ...
asjndna999's user avatar
0 votes
1 answer
49 views

Distribution of unbiased estimator given the sufficient statistic

Let T be a sufficient statistic for parameter a, and W be an unbiased estimator of a, then will the distribution of W|T always be independent of parameter a? I understand that T being sufficient for ...
Kcd's user avatar
  • 107
0 votes
0 answers
139 views

Subscripts for Expectations and variances in for estimators [duplicate]

Is there any significance for subscripts to E and Var? For example, the risk function of an estimator $\delta(\mathbf x)$ of $\theta$ in my book is: $$ R(\theta,\delta)=E_\theta[L(\theta,\delta(\...
Zachary Peskin's user avatar
5 votes
1 answer
320 views

How do these results show that $T(\mathbf{X})$ is an unbiased estimator of $E_\varphi[T(\mathbf{X})]$ that achieves the Cramer-Rao lower bound?

Let's say that $X_1, \dots, X_n$ has the joint distribution $f_\varphi(\mathbf{x})$ that belongs to the one-parameter exponential family $$f_\varphi(\mathbf{x}) = \exp{\left\{ c(\varphi) T(\mathbf{x}) ...
The Pointer's user avatar
  • 2,096
4 votes
1 answer
66 views

Does an estimator need to be unbiased in order to be sufficient?

I am reviewing some theoretical statistics content, and I was wondering if an estimator need to be unbiased in order to be sufficient? Is there any way to prove this? Thanks!
324's user avatar
  • 504
1 vote
1 answer
78 views

Did I correctly apply the factorisation theorem in this example?

Suppose that we have a density $f(x,\theta)=c(\theta)\psi(x)\unicode{x1D7D9}(x \in]\theta,\theta+1[)$ and the random variable $\mathbf{X}=(X_1,\ldots,X_n)$ are independently identically distributed ...
Hijaw's user avatar
  • 155
0 votes
0 answers
39 views

What's the maximum likelihood estimation of $\theta$ in this density? [duplicate]

Suppose we have a n-sample $X=(X_1,..,X_n)$ with a distribution $f(x,\theta)=exp(\theta - x)\unicode{x1D7D9}_{x \geq \theta}(x)$. Find the maximum likelihood estimator $T$ of $\theta$ and prove that $...
Hijaw's user avatar
  • 155
0 votes
1 answer
29 views

Proving sufficiency of a statistic using the expectation

I am blocked trying to solve the following question. I would appreciate if someone could give me a hint. Let $X_1,\ldots,X_n$ be $n$ independent random variables following a continuous uniform ...
QGM's user avatar
  • 109
1 vote
1 answer
3k views

Sufficient estimator for Bernoulli distribution using the likelihood function theorem for sufficiency

Let $(X_1,X_2)$ be a random sample of two iid random variables, $X_1\sim Ber(\theta),\theta\in (0,1)$. Use the following theorem to show that $\hat{\theta}=X_1+2X_2$ is sufficient. Likelihood theorem ...
stats19's user avatar
  • 61
1 vote
1 answer
242 views

nonexistence of a sufficient statistic

Let $X_1,X_2,\dots,X_n$ be a random sample from a $\Gamma(\theta,\theta)$ distribution. Then $$ \prod_{i=1}^n f(x_i;\theta) = \frac{1}{\Gamma(\theta)^n\theta^n}(\prod_{i=1}^n x_i)^{\theta-1}e^{-\frac{...
Tony B's user avatar
  • 220
2 votes
3 answers
131 views

"Magical" variance reduction problem

I recently came across this toy problem: You have two sticks of unknown lengths $a>b$ and a measuring device with constant variance $1$ that you can only use twice. How can you construct ...
Akababa's user avatar
  • 161
4 votes
1 answer
4k views

Invariance property

I am a bit confused regarding what exactly is the invariance property of sufficient estimators, consistent estimators and maximum likelihood estimators. As far as I know, Invariance property of ...
user233797's user avatar
7 votes
1 answer
426 views

Efficient Estimator from Insufficient Statistic

Suppose that I have a statistic $T(X)$, and I know for sure that it is not sufficient to estimate a parameter $\theta$. Is it still possible to have an estimator $\hat\theta(T(X))$ that is efficient (...
Cagdas Ozgenc's user avatar

15 30 50 per page