Skip to main content

All Questions

0 votes
0 answers
26 views

How to prove that the MLE of a uniform distribution is biased using the formula given below? [duplicate]

I've calculated the MLE of the uniform distribution on [0,theta] as maxi{Xi} but don't know how to prove it is biased. The formula I have learned to prove it is unbiased is E(θ^)-θ=0. Was stuck on how ...
meow's user avatar
  • 1
0 votes
1 answer
68 views

Definition of cost function and likelihood: how one appears from the other

I was reading this paper and the book "Introduction to Quantum State Estimation" by Yong Siah Teo and I am facing some issues trying to understand how the definition of the cost function ...
Dimitri's user avatar
  • 15
4 votes
1 answer
145 views

What do you think of this proof for Fisher information?

I want to prove This formula: The score function is basically the derivative of the maximum likelihood's log, so to get the information I make another derivative of that: $$ -E[∂/∂θ s(X;θ)] = -E[∂/∂θ ...
Programming Noob's user avatar
1 vote
1 answer
88 views

Finding the maximum likelihood solution corresponds to finding the root of a regression function. How?

Given a pair of RVs $z,\theta$ governed by a joint distribution $p(z,\theta)$. Conditional expectation of $z$ given $\theta$ defines a deterministic function (called as regression functions) $f(\theta)...
ironhide012's user avatar
1 vote
1 answer
59 views

Issue with Casella&Berger derivation of EM likelihood equality

In the explanation of the EM (Expectation maximization) algorithm p.328 in the book "Statistical inference" by G. Casella and R. Berger, 2nd edition, they present the following: $\mathbf{Y} =...
DancingIceCream's user avatar
1 vote
2 answers
5k views

Variance of MLE poisson distribution

I am working on problems related to finding MLE from Mathematical Statistics with Applications, 7th Edition - Wackerly. Below is the exercise 9.80 that I'm a bit confused over. My concern is mostly ...
OLGJ's user avatar
  • 337
2 votes
1 answer
88 views

Bias of MLE scales with $1/N$?

I was reading this paper (link) and it gave me some confusion. $P(r|\theta)$ is a distribution that generates sample $r$ based on some Poisson distribution, whose mean and variance are defined as some ...
CWC's user avatar
  • 281
2 votes
1 answer
343 views

What does "Expectation with respect to true unknown parameter" mean?

I am trying to study the asymptotic properties of MLE, but I am having trouble understanding an expression that seems to be consistently used in all lecture notes available online (page 93,page 18,...
CWC's user avatar
  • 281
3 votes
1 answer
711 views

Expectation for the MLE for a Uniform Discrete Random Variable

$\newcommand{\szdp}[1]{\!\left(#1\right)} \newcommand{\szdb}[1]{\!\left[#1\right]}$ Problem Statement: Suppose that $n$ integers are drawn at random and with replacement from the integers $1,2,\dots,N....
Adrian Keister's user avatar
3 votes
0 answers
431 views

Information Matrix for Conditional Likelihood

I am studying the MLE theory on my own and I am confused by the difference between the fisher information matrix for the full sample and for one observation, when it comes to conditional likelihood. ...
Count's user avatar
  • 1,389
10 votes
1 answer
3k views

Why is the observed Fisher information defined as the Hessian of the log-likelihood?

In an MLE setting with probability density function $f(X, \theta)$, the (expected) Fisher information is usually defined as the covariance matrix of the fisher score, i.e. $$ I(\theta) = E_\theta \...
flhp's user avatar
  • 101
1 vote
0 answers
47 views

Expected value of unbiased estimator of $\sigma$ in binomial sum

Suppose that $Y_1, Y_2, \dots, Y_r$ are random independent variable such as $Y_i \sim B(m_i, \pi)$, the idea first is to find $\hat{\pi}$ which is the maximum likelihood estimator an use it to find ...
jcaliz's user avatar
  • 121
1 vote
0 answers
126 views

Unbiased estimator for a parameter from a transformed distribution

I am solving an exercise in which I have to show that a certain estimator is unbiased for a given parameter. However, after a couple lines of computation I got stuck in the following scenario: $$ \...
bbublue's user avatar
  • 37
5 votes
1 answer
6k views

Difference between the expectation of x bar squared and the expectation of x squared

I am trying to understand the derivation of the expectation of the maximum likelihood (MLE) of variance, however I am confused as to what the difference is between $\bar{x}$ and $x$. Below you find ...
phil's user avatar
  • 341
2 votes
1 answer
958 views

Why is the expected value of variance different than the expected value of Maximum Likelihood variance?

The expected value of variance is: The expected value of the maximum likelihood of variance is: $E\left[\frac{1}{N}\sum_{n=1}^{N}(X-\mu_{ml})^{2})\right] = \frac{N-1}{N}\sigma ^{2} $ Why does ...
phil's user avatar
  • 341

15 30 50 per page