TransWikia.com

Properties of Legendre/Cramer's transformation of the moment generating function

Mathematics Asked on December 10, 2021

Let $X in L^1$ be a random variable on some probability space, define $M(theta) equiv E(e^{theta X})$ as its moment generating function and let $D(M) equiv {theta in mathbb{R} : M(theta) < infty}$. I’m reading a really toxic set of lecture notes which has the following lemma as an exercise and I cant find the same theorem online. I’m having trouble proving bullet point (iii) in particular:

enter image description here


My attempt at a proof follows. You may ignore bullet point (i) and (ii) if you wish, but they may be helpful:

For ease in notation I define $f_a(theta) equiv atheta – log(M(theta))$

Bullet point (i): Note $f_a(0) = 0 leq sup_theta f_a(theta) = I(a)$, convexity follows easily since $forall a, b in mathbb{R}$, $$(lambda a + (1-lambda)b)theta – log (M(theta)) = lambda (a theta – log(M(theta))) + (1-lambda)(btheta – log (M(theta))) leq lambda I(a) + (1-lambda)I(b)$$ and we can take the supremum of the LHS over $theta$.

We also have, by the convexity of $-log(x)$ and Jensen’s inequality, $$-theta mu = E(-log(e^{theta X})) ge -log(E(e^{theta X})) = -log (M(theta))$$ Rearranging gives $f_mu(theta) leq 0$ for all $theta$ (trivially for $theta$ such that the mgf is infinite) so that the supremum $I(mu) leq 0$. Lastly, we see that for each $b leq a leq mu$, we have the existence of a $lambda in [0,1]$ so that $a = lambda b + (1-lambda)mu$. From convexity, non-negativity and the fact that $I(mu) = 0$, it is immediate that $I(a) leq lambda I(b) leq I(b)$. The identical set of ideas works for $mu leq a leq b$

Bullet point (ii): For $a ge mu$, and any $theta < 0$, we have trivially $f_a(theta) leq f_mu(theta) leq 0$ from bullet point (i). The conclusion is trivial from here, and a similar set of inequalities is true for the other case.

Bullet point (iii): This is where I am lost. I know that $M'(theta)$ exists $forall theta in mathbb{R}$ since $X in L^1$ (for those who are unaware of how to prove this, simply find an appropriate dominating function for the difference quotient and apply dominated convergence), and moreover $M'(theta) = E(Xe^{theta X})$. So for the last part of (iii), we can note that $f_a(theta)$ is concave in $theta$ and differentiable so that any critical point is a global maximum. Thus, it suffices to find $theta$ so that (just take the derivative and set to 0) $$a – frac{M'(theta)}{M(theta)} = 0$$

This would actually solve the whole question, as it will show that we have a convex function that is finite at every point in $mathbb{R}$ (and thus finite on any open interval $(a,b)$) and is thus continuous (see here for example for a proof of this fact: Proof of "every convex function is continuous"). I just need help with the existence of such a $theta$. Please help if you can!

(As a sidenote, the fact that this is an exercise in a set of lecture notes is genuinely utterly cruel)

One Answer

Here is the proof that for any given $a$, if we define

$$g(theta) = frac{mathbb{E_mu}[X e^{theta X}] }{ mathbb{E_mu} [e^{theta X} ] }$$

there exists $theta^*$ s.t. $g(theta^*) = a$.

First, here is the intuition: You are given that X ~ $mu$ and $mu$ has full support on $mathbb{R}$. You can consider $a$ as the expectation of $X$ under the measure $mu_theta$ where :

$$ d mu_theta = frac{e^{theta x}}{mathbb{E_mu}[e^{theta X}] } d mu $$

First, note that this measure is well defined for all $theta$ since your MGF is finite for all $theta$. As $theta$ grows, this measure puts more and more mass on larger values, so eventually, under this measure, the expectation will arbitrarily grow. (Provided that $mu$ puts mass on these large values, which is how we will use that full support condition) We can play the same game with negative $theta$s, and we will show that you can get $g(theta)$ to be arbitrarily large and small. Since $g(theta)$ is continuous, intermediate value theorem will give you the result.

Now, let's do it rigorously:

Let $M$ be large. We know that $mu(X geq M) = epsilon$ for some $epsilon > 0$ since $X$ has full support on $mathbb{R}$.

We also have that $$lim_{theta rightarrow infty} mathbb{E_mu}[X e^{theta (X-M)} 1_{X < M} ] = 0$$ by dominated convergence theorem, so there exists a $theta_1$ s.t. $|mathbb{E_mu}[X e^{theta_1 (X-M)} 1_{X < M} ]| < epsilon$

We also have by a similar argument a $theta_2$ s.t. $|mathbb{E_mu}[ e^{theta_2 (X-M)} 1_{X < M} ]| < epsilon$

In what follows, let $theta = max(theta_1, theta_2)$

$$ frac{mathbb{E_mu}[X e^{theta X}] }{ mathbb{E_mu} [e^{theta X} ] } = frac{mathbb{E_mu}[X e^{theta X} 1_{X geq M} ] }{ mathbb{E_mu} [e^{theta X} ] } + frac{mathbb{E_mu}[X e^{theta X} 1_{X < M} ]}{ mathbb{E_mu} [e^{theta X} ] }$$

For the first term, we have that $frac{mathbb{E_mu}[X e^{theta X} 1_{X geq M} ] }{ mathbb{E_mu} [e^{theta X} ] } geq frac{ M mathbb{E_mu}[ e^{theta X} 1_{X geq M} ] }{ mathbb{E_mu} [e^{theta X} ] } = frac{ M mathbb{E_mu}[ e^{theta X} 1_{X geq M} ] }{ mathbb{E_mu}[ e^{theta X} 1_{X geq M} ] + mathbb{E_mu}[ e^{theta X} 1_{X < M} ] }$

Now, let's divide top and bottom by $mathbb{E_mu}[ e^{theta X} 1_{X geq M} ]$. Note that we can do this because $mu$ has full support. We get

$$frac{M}{1 + frac{mathbb{E_mu}[ e^{theta X} 1_{X < M} ]}{mathbb{E_mu}[ e^{theta X} 1_{X geq M} ]}} = frac{M}{1 + frac{mathbb{E_mu}[ e^{theta (X - M) } 1_{X < M} ]}{mathbb{E_mu}[ e^{theta (X - M)} 1_{X geq M} ]}} > frac{M}{2}$$.

where for the last inequality we note that $mathbb{E_mu}[ e^{theta (X - M)} 1_{X geq M} ] geq epsilon$ since $e^{theta (X - M)} geq 1$ when $X geq M$.

The second term is thankfully easier; we divide top and bottom by $e^{theta M}$

$$frac{mathbb{E_mu}[X e^{theta (X - M)} 1_{X < M} ]}{ mathbb{E_mu} [e^{theta (X-M)} ] } geq -1$$

where again we invoke the fact $mathbb{E_mu} [e^{theta (X-M)} ] > epsilon$.

Now, we are done, since we showed that for an arbitrary $M$, there exists a $theta$ s.t. $g(theta) > M/2 - 1$. We can play the same game with $-M$ which gives us another $theta'$ s.t. $g(theta) < -M/2 + 1$, which completes the proof by continuity of $g$ and intermediate value theorem.

Answered by E-A on December 10, 2021

Add your own answers!

Ask a Question

Get help from others!

© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP