Skip to main content

All Questions

28 questions with no upvoted or accepted answers
2 votes
0 answers
33 views

Lower bound of $\frac{\|(\mathbf X \otimes \mathbf X^\top)\theta\|_2^2}{np}$

According to Theorem 7.16 of High-Dimensional Statistics: A Non-Asymptotic Viewpoint (M. Wainwright, 2019), we know that for $\mathbf X\in\mathbb R^{n\times p}, X_{ij}\overset{iid}{\sim}N(0,1),$ there ...
Jasper Cha's user avatar
2 votes
0 answers
92 views

Probabilistic bound on difference of Lipschitz random function

I am currently facing the following problem : Let $(X_1,Z_1),\ldots,(X_n,Z_n)$ be $n$ i.i.d. sample points from some distribution $p$ supported on $\mathcal X\times\{-1,1\}$ where $\mathcal X\subseteq ...
Stratos supports the strike's user avatar
2 votes
0 answers
212 views

Hoeffding's Inequality Assumptions

I'm looking for the assumptions of the Hoeffding's inequality to check it is applicable to my problem. So far the only assumptions I can find are the variables $Z_i$ are IID and bounded. However, Im ...
curiouscat22's user avatar
2 votes
0 answers
70 views

Show that $\operatorname{Pr}(Z-X \geq 0)$ converges to one

Suppose that $V_i$, for $i \in \mathbb{N}$, are i.i.d. standard normal random variables and $Y_i = \sum_{k=1}^i V_k$ for $i \in \mathbb{N}$ with $Y_0 = 0$. Let $X_n = (\sum_{i=1}^n V_i Y_{i-1})^2 Y_n^...
KRL's user avatar
  • 1,180
1 vote
0 answers
27 views

Showing upper and lower Bayesian method of survival function

\begin{equation} \begin{split} S(t) = \frac{1}{\int_{0}^{1}\prod_{m=1}^M\left( \prod_{i=1}^{n_m} (t_{mi}+yx)^{\beta-1}\right)y^{c-1}(1-y)^{d-1}\frac{ \Gamma(\sum_{m=1}^{M}n_m+a)}{\left[\sum_{m=1}^{...
Mmmm's user avatar
  • 11
1 vote
0 answers
17 views

Calculation of lower range confidence interval

Question We observe $x$, the maximum of $n$ values in a random sample from the uniform distribution between $0$ and $c$, where $c > 0$. Find an exact lower range $100(1 - \alpha)\%$ confidence ...
Ethan Mark's user avatar
  • 2,187
1 vote
0 answers
48 views

Product of random variables greater than dependent sum of random variables

Let $X := {1\dots m}$ be the set of indexes corresponding to the elements of the vector of i.i.d. random variables $w:=(w_1, \dots, w_m)$. Let there be the subsets $Y, Z \subseteq X$ which may or may ...
Scriddie's user avatar
  • 221
1 vote
0 answers
27 views

Tight bounds for the expected maximum value of k IID Binomial(n, p) random variables

What is the tightest lower and upper bound for the expected maximum value of k IID Binomial(n, p) random variables I tried to derive it : $$Pr[max \leq C] = (\sum_{i = 0}^C {n \choose i}p^i(1 - p)^i)^...
Goli Emami's user avatar
1 vote
0 answers
90 views

Proving the set where probability density function becomes infinite is bounded

For a continuous random variable $X$, with probability density function $p_X(x)$, it is known that there exists a $p_{min} > 0$ such that $p_X(x) \geq p_{min} \forall x \in X$. Also, I know that $X$...
Janne's user avatar
  • 11
1 vote
0 answers
48 views

Upper Bound for Moments for Product of Sample Means

I have a question about the upper bound of the following moment. Suppose that $(A_1, e_1),\ldots, (A_n, e_n)$ are i.i.d. with $E(e_i)=0$. I am wondering if we have the bound $$E\bigg(\bigg\|\frac{\...
beginner's user avatar
1 vote
0 answers
102 views

Concentration bound for the distribution of the difference of two random variables

If we use $\Rightarrow$ to represent convergence in distribution and suppose that $X_n \Rightarrow N(0,\sigma_1)$ and $Y_n \Rightarrow N(0,\sigma_2)$, and $X_n$ and $Y_n$ are independent, then we all ...
lmz's user avatar
  • 11
1 vote
0 answers
20 views

Use the statistical process control method to find the indicated control limits.

Question : The table gives $10$ samples of three measurements, made during a production run. Use the statistical process control method to find the indicated control limits. Using $k_2=2.568$ and $...
Hayden Wilcox's user avatar
1 vote
0 answers
46 views

Generalization Bounds

Given the loss function $L(\hat{y},y)$ the generalization error is defined as $$R(h) = \underset{(x,y)\sim D}{\mathrm{E}}[L(h(x),y)]$$ the empirical error is defined as $$\hat R(h) = \frac{1}{m}\...
Swornim Baral's user avatar
1 vote
0 answers
52 views

Does logistic regression not fulfill an inequality required for Wilks' Theorem or am I missing something?

The required inequality: Wilks' Theorem is given in the source below as Theorem 12.4.2, p. 515. Before stating the inequality, some definitions are needed: Let $Z_1, \dots, Z_n$ be i.i.d. according to ...
MathStudent's user avatar
1 vote
0 answers
20 views

On the difference between the main effect in a one-factor and a two-factor regression

This question was asked on Cross Validated where it received little attention and no comments or answers, but as it is purely mathematically oriented it may well be more suitable here. Consider a ...
Arnaud Mortier's user avatar

15 30 50 per page