Search Results
Search type | Search syntax |
---|---|
Tags | [tag] |
Exact | "words here" |
Author |
user:1234 user:me (yours) |
Score |
score:3 (3+) score:0 (none) |
Answers |
answers:3 (3+) answers:0 (none) isaccepted:yes hasaccepted:no inquestion:1234 |
Views | views:250 |
Code | code:"if (foo != bar)" |
Sections |
title:apples body:"apples oranges" |
URL | url:"*.example.com" |
Saves | in:saves |
Status |
closed:yes duplicate:no migrated:no wiki:no |
Types |
is:question is:answer |
Exclude |
-[tag] -apples |
For more details on advanced search visit our help page |
Results tagged with maximum-likelihood
Search options answers only
not deleted
user 526995
For questions that use the method of maximum likelihood for estimating the parameters of a statistical model with given data.
0
votes
Maximum Likelihood Estimator Independent Exponentials
Finding the MLE: You should be more careful with your sigma-notation in your working, by bracketing the terms in the sum. The simplest way to express the score log-likelihood here is in terms of the …
0
votes
Maximum likelihood estimator of $b$, given $n$ draws from the minimum distribution of two va...
Your answer is not correct. From your (correct) specification of the density for the minimum value, the log-likelihood is:
$$\ell_{\boldsymbol{x}} (b) = \sum_{i=1}^n \ln(b - x_i) - 2n \ln(b) \quad \ …
2
votes
How to find the MLE of the parameters of an inverse Gaussian distribution?
Consider the simpler case of $X_1,...,X_n \sim \text{IID InvN}(\mu, \lambda)$, which gives you the sampling density:
$$f(\mathbf{x} | \mu, \lambda) = \prod_{i=1}^n \Big( \frac{\lambda}{2 \pi x_i^3} \B …
0
votes
Estimating Regression Coefficients where parameters include linear combinations
Maximum-likelihood estimators are invariant under functional transformations, and so if $\boldsymbol{\alpha} = \boldsymbol{\beta} \boldsymbol{A}$ is a new parameterisation for your regression then $\h …
0
votes
Asymptotic Maxwell MLE distribution
There is no need to use asymptotic approximations here, since we can get the exact distribution of the MLE. The distribution you didn't recognise is $Y = X^2 \sim \text{Gamma}(\text{Shape} =\tfrac{3} …
1
vote
Maximum likelihood estimator for translated uniform distribution
Your answer is correct, and it is simple to verify using ordinary calculus methods. It is usually simpler to express this with the log-likelihood rather than the likelihood, since it is easier to dif …
0
votes
BLUEs of angles of triangle
As your question stands you have independent observations $Y_i \sim \text{N}(\beta_i, \sigma^2)$. There is no relevance to the contextual statement that these are angles of a triangle, since you have …
1
vote
Accepted
Finding maximum likelihood estimator of two unknowns.
The approach you have used is fine, and what you are essentially doing is getting MLE equations for each parameter in terms of the other one, and then solving these as simultaneous equations to get th …
1
vote
Determining maximum likelihood estimator for scaled beta distribution
In problems like these you can still use ordinary calculus methods; you have a monotonic likelihood function with the MLE occurring at a boundary point. This gives rise to a biased estimator of the p …
6
votes
Accepted
Maximum Likelihood Estimation for Zero-inflated Poisson distribution
The book you have referenced uses some general theory about zero-inflated distributions (i.e., the application of some results that are not specific to the Poisson case). I have been unable to replic …
0
votes
Defects of Least square regression in some textbooks
Since you have stated that $y = ax+b$, it is unclear in your question where the errors come into your model. However, if you are saying that there is some error in the measurement of $x$ and $y$ then …
1
vote
Accepted
MLE of AR(2) time series model
The conditional MLE for an auto-regressive model can be solved using standard linear regression theory. It is easy to do this in matrix form and use the standard formula for the coefficient estimator …
1
vote
Accepted
Find the maximum likelihood estimator of b (Regression coefficient)
In this case your log-likelihood function is:
$$\ell (b) = - \sum_{i=1}^n | y_i - b x_i |.$$
Since your $x_i$ values are either negative or positive one, we can let $\mathscr{X} \equiv \{ i = 1,..., …
1
vote
Likelihood Ratio Test for Exponential Distribution with a Limited Parameter Space
In the case where you observe $\bar{x}_n \leqslant 1$ you have test statistic $t = 0$ and the corresponding p-value for your test is $p=1$. This means that there is no evidence to reject the null (i. …
0
votes
If best unbiased estimator exists then it's maximum likelihood estimator?
This cannot be true as a general theorem, since there are situations where the MLE is not unbiased, but other unbiased estimators exist. For example, this occurs when estimating the parameter $\theta …