Talk:Characteristic function (probability theory)

Discussion…
I think the formula for calculating the n-th moment based on the characteristic function was not correct. It was:

$$E(X^n)=-i^n \varphi^{(n)}_X(0)$$.

I changed it to

$$E(X^n)=(-i)^n \varphi^{(n)}_X(0)$$.

Please compare with.


 * In fact, neither of those are correct. I will correct both occurrences of the formula in the article to $$E(X^n)=i^n \varphi^{(n)}_X(0)$$.  —Preceding unsigned comment added by 188.36.27.144 (talk) 14:00, 11 October 2009 (UTC)


 * Why?! We have (assuming analyticity),
 * $$ \phi(t) = E \sum \frac{(itX)^n}{n!} = \sum \frac{t^n}{n!} i^n E X^n;$$
 * $$ \phi^{(n)}(0) = i^n E X^n;$$
 * $$ E X^n = \frac1{i^n} \phi^{(n)}(0) = (-i)^n \phi^{(n)}(0).$$
 * Right? Boris Tsirelson (talk) 16:45, 11 October 2009 (UTC)

"Different change of parameter"
What does this phrase mean?: “which is essentially a different change of parameter”. —Preceding unsigned comment added by 130.207.104.54 (talk) 15:56, 20 August 2009 (UTC)


 * It means that one function is just a rescaled version of the other:
 * $$\varphi_2(t) = \varphi_1(-2\pi t)$$
 * Second definition is more standard in the Fourier analysis, while former is generally used in statistical sciences. ...  st pasha  » talk » 10:50, 21 August 2009 (UTC)

Inversion theorem
I have a bit of a different definition for the inversion theorem that seems to contradict the one in this article. From "probability and random processes" by Grimmet I have:


 * $$ \overline{F}(x) = \frac{1}{2}[F(x) + \lim_{y \to x^-}F(y)] $$

then
 * $$\overline{F}_X(y) - \overline{F}_X(x) = \lim_{\tau \to +\infty} \frac{1} {2\pi}

\int_{-\tau}^{+\tau} \frac{e^{-itx} - e^{-ity}} {it}\, \varphi_X(t)\, dt.$$ --Faisel Gulamhussein 22:17, 20 October 2006 (UTC)

Probably you're correct. I don't think there is a contradiction exactly. Just many authors assume their random variables have a density, if this is the case the statements would be the same. Thenub314 13:07, 21 October 2006 (UTC)

Would someone like to write a brief explanation of the meaning of a characteristic function, for those of us without such strong statistics and mathematics backgrounds? I have no idea of its concept or application. An example that appeals to intuition would be great, thanks! BenWilliamson 10:28, 21 February 2007 (UTC)

Bochner-Khinchin theorem
There is something strange as it is formulated.

The would-be characteristic function

$$ \varphi(t) = e^{- c t^4} $$ satisfies all the requested properties but it implies that the corresponding distribution has vanishing second moment

$$ E[X^2] \propto \varphi''(0)=0 $$

--MagnusPI (talk) 09:25, 24 April 2008 (UTC)

I believe the problem lies in condition: " $$ \varphi(t) = e^{- c t^4} $$ is a positive definite function". I suggest that this doesn't hold, and there are results that indirectly prove it isn't ... but I don't know of a direct approach to showing this. Note that "positive definite function" is a non-straightforward math thing if you are not familiar with it. Melcombe (talk) 09:34, 25 April 2008 (UTC)

On any compact $$ \varphi(t) = e^{- c t^4} $$ is positive definite, the only points where it is not are $$ \pm \infty $$ but in this case I do not understand why the characteristic function for some Levy stable $$ \varphi(t)= e^{ - | c t |^\alpha } $$ can be accepted since they show the same behaviour at $$ \infty $$

--MagnusPI (talk) 08:44, 28 April 2008 (UTC)


 * I suggest you follow the link to positive definite function to see what is required .. it is not just $$ \varphi(t) > 0 .$$ Melcombe (talk) 08:56, 28 April 2008 (UTC)

Interesting but this is not what a physicist would call a positive definite function. I will therefore add a note to the main page

--MagnusPI (talk) 08:55, 29 April 2008 (UTC)

History of Characteristic Functions
Historia Matematica has a nice little thread on the history of characteristic functions which could be used to give some background on where c.f.'s came from. --Michael Stone 19:22, 10 April 2007 (UTC)

Please check your work.

$$E(X^n)=i^n\varphi^{(n)}_X (0)$$.

Inverse fourier transform
The article mentions that you can calculate the characteristic function taking the conjugate of the Fourier transform of the pdf, but isn't this simply the inverse Fourier transform?... This should be made more clear. We tend to think about transforming using only the "direct" transform first, and then transforming back again, but you can perfectly say that you can find the characteristic function taking the inverse Fourier transform, then come back again taking the direct Fourier transform...

It's like the pdf is already the "transform", and then the characteristic function is the "original" function, that we could see as a signal to be transformed...

The idea is just that we save some words using the idea of "inverse fourier transform" instead of "conjugate of the transform", and also help to see the inverse transform as something as natural as the "direct" transform. No transform domain is intrinsically better! there is no reason to insist in seeing the pdf as the correlate of a signal to be transformed by the direct fourier transform, and not the inverse one... —Preceding unsigned comment added by Nwerneck (talk • contribs) 22:53, 17 November 2007 (UTC)

Pathological examples
The article gives the impression that every random variable can be described by a characteristic function; I am almost sure (no, that's too strong to say in a statistics page - let me rephrase it: I have more than 95% confidence) that a pathological example can be designed that does not have a characteristic function... Something along the lines of the perforated interval in page standard probability space. Darth Albmont (talk) 19:06, 18 November 2008 (UTC)
 * I suggest you see page 11 of the Lukacs reference. He cites a theorem of Cramer (1946) which he says ensures the existence of the cf for "every distribution function". The previous context makes it clear that this includes even dfs with singular components. See also page 20, 36 and 64 for other results on the relation between the cf and whether the df has a singular component. Melcombe (talk) 10:01, 19 November 2008 (UTC)
 * Does every random variable defined in the real numbers have a distribution function? Darth Albmont (talk) 12:25, 19 November 2008 (UTC)
 * You linked to density function but wrote distribution function; they are not equivalent. Every real valued RV has a distribution function; not all have density functions (e.g., Cantor distribution).  Your original question was about characteristic functions; they always exist for any distribution with a Lebesgue measure due to the dominated convergence theorem.  I cannot say for sure about distributions with more pathological measures, but I suspect it's true there too.  I cleaned up the wikilinks in your original post. Baccyak4H (Yak!) 15:02, 19 November 2008 (UTC)
 * You linked to density function but wrote distribution function; they are not equivalent - Baccyak4H, yes, but the page distribution function is about physics, and it gives both cumulative and density as the probability concept. I don't know if there are pathological real valued random variables so pathological that it's not possible to compute its characteristic function - but let's try a counting argument. Let X be any set that is dense in [0,1]; is it possible to define a real-valued random variable whose sample space is X? If we accept the Axiom of Choice (the psychomath's favourite tool in creating monsters) and the Continuum hypothesis, then the answer is yes: either X is countable or has the cardinality of the continuum; in the first case, it's trivial, in the second, well-order both X and [0,1], sample uniformly from [0,1] and get back to X using the bijection (probably there's no need to use the continuum hypothesis to create this monster). So, by counting, there are 2c real-valued random variables. However, since every characteristic function is continuous, there are only c characteristic functions. Unless I made some big mistake, I think I proved that not all real-valued random variables have a characteristic function (in fact, I guess I showed that most of them don't - the problem with pathologies is that usually pathologies are the rule). Darth Albmont (talk) 16:33, 19 November 2008 (UTC)


 * The characteristic function is just the Fourier transform of the distribution. (Thus it is really about distribution rather than random variable itself.) It is well-defined always. At least if we treat a distribution as a probability measure on the Borel sigma-field (be it completed or not). No place for pathologies here. Boris Tsirelson (talk) 20:17, 23 March 2009 (UTC)


 * Characteristic function is an integral $$\phi_X(t) = \mathrm{E}\,e^{itX} = \int e^{itx}dF_X(x)$$. Since expectation operator exists for any random variable, and the corresponding integral is absolutely integrable for real values of t: $$\left| \int e^{itx}dF_X(x) \right| \leq \int |e^{itx}|dF_X(x) = \int dF_X(x) = 1$$, then characteristic function is well-defined for any random variable, even a "pathological one". Stpasha (talk) 21:20, 16 June 2009 (UTC)

Matrix-valued random variables
In the above-named section, what the dimensions of the matrix T in relation to those of X? Are the dimensions the obvious ones or are other sizes used occasionally? Melcombe (talk) 15:45, 9 December 2008 (UTC)
 * Hi Melcome. I've added some material to the page; hope it is clear.  Let us know if not.  Best wishes, Robinh (talk) 08:05, 10 December 2008 (UTC)


 * Yes it is clear. But can I ask if this formulation is used in practice anywhere? It seems that if T were replaced by T-transpose then, not only would the formula be directly comparable to the vector case, but both the following would apply
 * each Xij would be associated with Tij rather than Tji,
 * the formulas would be equivalent to the vector case if both X and T were strung out into vectors using the usual vec operator.
 * Melcombe (talk) 09:41, 10 December 2008 (UTC)
 * Hi. Well you could define a function like the characteristic function that works as you specify.   But this would not take advantage of the structure present in matrices.  The characteristic function as specified in the article is the standard, which crops up all over the place, at least in distributions such as the Wishart distribution.  Best wishes, Robinh (talk) 13:16, 10 December 2008 (UTC)

Moment generating function
I would like to have any mention of the moment-generating function (mgf) removed from the "definition" section. The concept of characteristic function can and should be defined without any reference to mgf; in fact such reference is only confusing, because unlike cf, mgf may not always exist, or may be unbounded, etc.

The reference itself is quite dubious as well. The expression $$M_{iX}(t)$$ denotes an mgf of a complex-valued random variable. I'm not entirely sure how mgf for a complex-valued r.v. is supposed to be defined, but most likely the argument t will be complex as well. The expression $$M_X(it)$$ is not correct. Function $$M_X:\mathbb{R}\to\mathbb{R}$$ is defined over a domain of real numbers, it cannot be evaluated at an imaginary point (at least not from the point of view of strict mathematics). Of course if function $$M_X$$ is given by some simple formula then we can always plug in $$it$$ instead of $$t$$ and hopefully get a meaningful result. But what if there is no simple formula? What if $$M_X$$ is defined graphically for example. It is difficult to interpret $$M_X(it)$$ without going back to definition of $$M_X$$, but then such interpretation turns into a tautology.

Anyways, i believe that mgf is nothing more than a "related concept", and thus should be mentioned in the corresponding section at the end of the article.

Stpasha (talk) 11:24, 17 June 2009 (UTC)


 * Some of the old notation involving was wrong and I have taken it out. Many people do define mgf's as allowing complex arguments. The similarity in the formulae for cf's and mgf's means that mgf's have to be mentioned early on to prevent readers wondering why they are not mentioned. On there is no reason why they should not be mentioned more than once if there is an increase in technicality towards the end of the article. It is wrong to think that either cf's or mgf's are only defined for real-valued arguments and I have tried to introduce something about this. Melcombe (talk) 13:27, 17 June 2009 (UTC)


 * I do agree that moment-generating functions are treated in many current textbooks, although the reasons for that are either historical (mgf was invented earlier than cf), or the authors are unwilling to go into the area of complex analysis.
 * However, mgf does not bring anything new to the theory compared to cf. Surely mgf can be used to calculate moments, or to formulate a convergence theorem. But cf already does that -- we can calculate moments by differentiating cf and we have Lévy continuity theorem. On the other hand mgf has quite a few drawbacks compared to cf.
 * And it is wrong to think that cf's or mgf's are defined for unreal-valued arguments. In fact, if we try to plug complex-valued t into cf, we will arrive at something called analytic characteristic function (see Lukacs, ch.7), which loses some of the nice properties of cf (for example existence), but gains some extra useful properties (ridge property, existence of ∞ number of moments, sometimes entire-ness). It is when we consider analytic cf, the connection with mgf becomes well-defined. However analytic cfs are different from regular cfs, they deserve their own section, and should not be implied in the definition of regular cf. Stpasha (talk) 21:14, 17 June 2009 (UTC)
 * Remember that this is not a text book. It needs to be useful to many types of people and to include what is notable about the topic and related topics. the connection to mgf's is definitely that. It is wrong to say that "to plug complex-valued t into cf" will arrive at analytic cfs, ... it arrives at cfs that may or may not be analytic over regions to be defined, as per Lukacs. Just because you like thing think of real 't' only is no reason for the article to give the impression that only that case is important. And considering that the major work on characteristic functions defines them for complex t it needs to be included in the main definition. Also, before you add even more unexplained maths in a prvate notation into the initial section, read WP:LEDE. Melcombe (talk) 08:50, 18 June 2009 (UTC)

Standard normal
Does anybody know if standard normal N(0,1) is the only random variable whose characteristic function coincides with its pdf? Then it would be a fixed point of the Fourier transformation; wonder what this would imply. Stpasha (talk) 18:35, 30 June 2009 (UTC)


 * Well, you have to be a little careful N(0,1) is not quite the same as its characteristic function, depending on how you define the characteristic function. The characteristic function defined here is missing the normalizing factor involving 2π.  It turns out there are many distributions whose characteristic functions are themselves except for a normalizing factor in this way.  Thenub314 (talk) 06:53, 1 July 2009 (UTC)


 * Yep, my bad. This should use alternative definition of characteristic function φ=E[e-2πitX] (because only then Fourier transform is unitary), and then fixed point distribution will be N(0, (2π)-1) (seeing as we need fX(0)=1). But still what other "many distribution whose cf's are themselves except for a normalizing factor"? // Stpasha (talk) 07:56, 1 July 2009 (UTC)


 * The unitary operator of Fourier transform has four eigenvalues: 1, i, –1. –i (and no continuous spectrum). The corresponding eigenfunctions are Hermite polynomials times the normal density. Their eigenvalues are: 1, i, –1. –i, 1, i, –1. –i, 1, i, –1. –i, ... Thus, you may add to the normal density the fourth Hermite polynomial (times the normal density) with a small coefficient of such sign that for large x the addition is positive. This way you get an example you seek. Boris Tsirelson (talk) 12:47, 1 July 2009 (UTC)


 * Another way is to consider any Schwartz function &fnof; and take $$f+\mathcal{F}(f)+\mathcal{F}^2(f)+\mathcal{F}^3(f)$$, then when you apply a Fourier transform to such a thing you get out what you started. Thenub314 (talk) 13:07, 1 July 2009 (UTC)


 * Yes, but we also want it to be positive (and integrable). Boris Tsirelson (talk) 18:40, 1 July 2009 (UTC)


 * I hope you'll agree integrability is handled by the fact that I said Schwartz. I should have said something about positivity, but to ensure that one simply needs to take a function whose Fourier transform is positive.  For completeness of this discussion here are a few examples aside the gaussian: e-; sech(x); &fnof;*&fnof; for any even positive Schwartz function &fnof;, and many others.  For the concerned reader I should mention why these examples and you sum the iterates of Fourier transforms is not necessarily the Gaussian, one easy way to see this it to look at the asymptotic as x → ∞, the first example will decay like 1/x2, the second will decay like e-x, for the third it of course depends on what you pick for &fnof;.  The example suggested by Boris is much better in the sense it is much more systematic, and much more complete.  I just thought this was an interesting way to look at it, and it perhaps generates some closed formulas that are not covered by Boris's argument. Thenub314 (talk) 09:27, 2 July 2009 (UTC)


 * Alright, so Hermite polynomials didn't really work; however density of the form $$(ax^4-\tfrac{3}{2\pi}ax^2+1)e^{-\pi x^2}$$ (where $$a \leq (4\pi/3)^2$$) indeed coincides with its own cf. Although it certainly cannot be expressed as a Hermite polynomial since they have integer coefficients.
 * I thought about Boris's example some more, and am now fairly sure that any function of the form
 * $$ \frac{(-1)^n}{2^n(2n-1)!!}H_{2n}(\sqrt{2\pi}x)e^{-\pi x^2}$$
 * (where H is a physicist's Hermite polynomial) is an eigenfunction of cf. transform. Although not an appropriate pdf, we can always add a normal distribution to it to ensure positiveness. // Stpasha (talk) 19:28, 1 July 2009 (UTC)


 * I am a bit confused by this last comment, we seem to have moved from showing there are a few any other then the Gaussian distribution to describing the space of all probability distributions with this property. This problem is a bit harder.  Here are some random thoughts about that.  First we are looking for probability measures whose Fourier transform is itself, but the Fourier transform of a probability measures is a bounded function, and the measure has integral one, so in fact our probability measure must have a bounded pdf.  This pdf, being bounded and in L1 must also be in L2. Following Boris's suggestion, we could consider a basis in L2 for the eigenspace associated to eigenvalue 1.  If you want to describe all elements of this eigenspace that are positive and integrable then it depends on what you mean by describe (aka why is it not enough to say take everything in the eigenspace that is positive and integrable.)  I suppose it could be useful to notice that we are looking for positive definite integrable functions in this space.  By a theorem of Bochner if the function is positive definite the characteristic function will be positive, since these functions are themselves under the Fourier transform, this turns into a condition for the function itself to be positive. What would really be nice is to have a condition on the coefficients of the expansion in this basis that guaranteed that resulting function would be positive, and integrable.  Unfortunately nothing springs to mind that would tell me when this was the case.  Thenub314 (talk) 09:27, 2 July 2009 (UTC)


 * I suppose we have that actually we might be in better shape than I first thought. We have that $$\tfrac{1}{4}(I+\mathcal{F}+\mathcal{F}^2+\mathcal{F}^3)$$ is a projection in L2 onto the functions that are their own Fourier transform. From the reasons above we are interesting in finding positive functions that are integrable with integral 1 (from being a probability measure), and are bounded (from being the Fourier transform of a probability measure).  All such functions are in L2, so take your favorite one and apply the projection.  We still can't categorize them exactly, but we know that doing this we hit all of them.  Here are some properties of all such functions &fnof; that might be interesting.
 * &fnof; must be positive, and have integral one since it defines a probability measure.
 * &fnof; must be bounded by 1, because it is the Fourier transform of a probability measure (itself).
 * &fnof;(0)=1 because its value at zero is equal to the integral of its Fourier transform, which is in turn the integral if itself.
 * &fnof;||p &le; 1 for all 1 &le; p &le; ∞ (interpolation).
 * &fnof; is uniformly continuous and tends to 0 as |x| → ∞, because it is the Fourier transform of an integrable function.
 * Applying our projection to the triangular function gives an example to show that &fnof; need not be differentiable, or decay faster then 1/x2
 * &fnof; must be positive definite because by Bochner's theorem a positive function is the Fourier transform of a measure if and only if it is positive definite and continuous.
 * &fnof; is real-valued, so it must be and even because its Fourier transform is real valued.
 * &fnof; must be supported on the entire real line.


 * The list could probably be extended. For example, I it probably follows from the uncertainty principle that the function also cannot decay too rapidly (e-x 4 is out).  On the other hand if the function is smooth it cannot decay too slowly either.  I guess this above paragraph goes under the "what would this imply" part of your original question. Thenub314 (talk) 11:28, 2 July 2009 (UTC)

This can't be right, surely?
The article says that, if phi is a characteristic function, then so is Im(phi). But one property all characteristic functions phi have is phi(0) = 1, so Im(phi)(o) = Im(phi (0)) = Im(1) = 0, which is not 1. So, surely, if phi is a characteristic function, then Im(phi) never is? I tried correcting this by removing Im(phi) from the list, but I never saved my edit. I previewed it, only to find out I'd messed up the code on all the other symbols for Re(phi) etc. Could someone who knows what they're doing with the code please remove Im(phi)? (Unless it actually does belong there, but for the life of me I cannot see how!) 90.206.183.244 (talk) 15:38, 6 July 2009 (UTC)


 * Thank you. You are right. I did. Boris Tsirelson (talk) 18:36, 6 July 2009 (UTC)

Properties - symmetry
Under "properties" the article states "the characteristic function of a symmetric random variable is real-valued and even". A similar remark is made in the caption for graph of the characteristic function of a uniform distribution. But is this really generally true, or true only if the random variable is symetric around zero. For example, a normal or Cauchy distribution with a non-zero mean is symetric, but it appears that the characteristic functions would retain complex values. But I admit that I am weak on complex math, so maybe those i's that seem to be there drop out. Rlendog (talk) 13:54, 21 August 2009 (UTC)
 * You are right. Thanks. I just added "around the origin". Boris Tsirelson (talk) 15:07, 21 August 2009 (UTC)

The same space, or the dual space?
"The argument of the characteristic function will always belong to the same space where random variable X takes values" (Sect. "Definition") — really? In principle, it belongs to the dual space. In practice, the given space usually is endowed with an inner product and so, may be treated as dual to itself. But not always; the distinction may be essential for random processes (because of infinite dimension). In fact, the distinction manifests itself already in finite dimension, if the given space is just an abstract linear space (with no preferred basis). Boris Tsirelson (talk) 17:42, 24 September 2009 (UTC)


 * You may be right, I haven’t seen a book which defines the cf for generic random elements. This sentence is just my attempt to describe those several definitions that follow. In most cases of course the dual space is isomorphic to the the “primal” space, so that for example if X&isin;R then the argument of cf is also t&isin;R. And if X&isin;Rk then also t&isin;Rk. In any case it may be just the matter of point of view: whether you want to see the product t′X as a linear map acting on variable X, or as an inner product (t, X) between two elements of the same space over the field of reals.
 * The case of cf of the stochastic process is trickier though. Instead of having X∈L² and t∈L² we have X arbitrary, and t belonging to some strange class of functions integrable in product with X. Maybe this is what the dual space is.  …  st pasha  »  18:32, 24 September 2009 (UTC)

So, if t is in the dual space, then which dual is it — algebraic or continuous? Of course the two duals coincide in finite dimensions, but suppose we want to consider the cf of an infinite-dimensional random element taking values in a Hilbert space H. Then the algebraic dual is"larger" than the original space, and they are not isomorphic. That would mean that the characteristic function has"more dimensions" than the distribution function. If we consider the continuous duals however, then by Riestz representation theorem this space is isomorphic to the domain of the random variable, and there isn’t much error in saying that t lies in the same space as X. //  st pasha  »  20:58, 26 September 2010 (UTC)


 * "which dual is it — algebraic or continuous?" — continuous (otherwise, non-measurable...)
 * "by Riestz representation theorem..." — and what if the original space is Lp, not just L2? The dual to Lp is some Lq, see Lp space. Boris Tsirelson (talk) 21:55, 26 September 2010 (UTC)
 * "I haven’t seen a book which defines the cf for generic random elements" — here is one: V.I. Bogachev, "Gaussian measures" (AMS); see Sect. 2.2. Boris Tsirelson (talk) 22:05, 26 September 2010 (UTC)

Notation
How about we denote the characteristic function with letter &chi; instead of &phi;? The reason being that &phi; is a reserved symbol for the pdf of the standard normal distribution. …  st pasha  »  00:04, 5 October 2009 (UTC)


 * And &chi; is reserved for the indicator function; and by the way, non-probabilists often call it characteristic function of a set! It is impossible to reserve every symbol once and for all. But if you see &chi; for the characteristic function in some textbooks, then we can follow these. Boris Tsirelson (talk) 09:42, 5 October 2009 (UTC)


 * Luckily we have an unambiguous notation for the characteristic function of a set: 1(A) or 1A. Now I don't know if any textbook actually uses &chi; for c.f., and besides it clashes with the chi-square distribution... As an alternative, we can use ϕ and φ to denote the c.f. and the standard normal pdf. Only need to decide which one is which (I personally vote for ϕ being the standard normal pdf, since it is more similar to Φ).  …  st pasha  »  22:34, 5 October 2009 (UTC)

How about providing a brief explanation of terms, such as t in e^it? That would be great to improve understanding of the cf for those who have not taken advanced courses in math. "and t∈R is the argument of the characteristic function" is not fully comprehensible. I can only conclude that this is a real number, a continuous quantity. Does it always changes from -infinity to +infinity? And what is the role of this argument?
 * Well, R denotes the real line, which means it goes from −∞ to +∞. Now, the characteristic function is a function, so it has the argument. That is, t is a point where you want to evaluate your function. If you want to find what φ(0) is, plug in t=0 in the formula: E[e0iX]. If you are looking for φ(5), you plug in t=5 and compute E[e5iX]. There isn't even an article for the word argument of a function, so i guess it is considered to be a trivial concept...  //  st pasha  »


 * Thanks for explanaions. Actually, I asked about the physical meaning of t. For example, I understand the meaning of p in, say, binomial distribution, but I cannot link t with something that is sensible. A cumulative distribution function or a probability mass function (probability density function) contain easily interpreted arguments, while the interpretation of t in cf is vague for me. Can you explain me this, please? —Preceding unsigned comment added by 87.119.246.143 (talk) 14:26, 25 May 2010 (UTC)


 * I haven't seen a book where the physical interpretation of t would have been discussed. However I think it's possible to think of it as some sort of “velocity”, borrowing from the Heisenberg uncertainty principle. For example, if location of a particle is described by a Gaussian r.v. with mean zero and variance &sigma;², then the pdf is proportional to e−x²/(2&sigma;²), whereas the c.f. will be equal to e-x²/2·&sigma;². Thus one curve has the “width” &sigma (this is the uncertainty in position of x), while the other has width 1/&sigma; (the “uncertainty” in “velocity” of x). The point being that the product of these two uncertainties can never be smaller than a certain positive constant. Ok, I successfully confused myself. The “speed” of a random variable???  //  st pasha  »  23:58, 25 May 2010 (UTC)

Another suggestion is to provide examples of derivation cf for, say, bernulli, binomial and normal distributions. Such examples are useful for quick learning of what cfs are. User:NoName (contribs) 17 May 2010 (UTC)
 * For Bernoulli this is simple: X takes on only two values: 0 and 1, with probabilities 1−p and p, so the expected value will be E[eitX] = (1−p)e0 + peit = 1 − p + peit. For a normal distribution, the derivation is pretty complicated since it requires the knowledge of complex analysis.  //  st pasha  »
 * A 'physical' interpretation for t isn't easy, or maybe not even possible, but as an ex-physicist myself i appreciate where you're coming from in seeking to gain intuition about what t "means" and i've pondered it myself a while back. It helps if you've come across Fourier transforms, as the characteristic function is the Fourier transform of the pdf. If you haven't, the most familiar physical example of a Fourier transform is the relationship between waveforms and frequency spectra, e.g. of sound. (In quantum mechanics wavefunctions for position & momentum, and also energy & time, are related by Fourier transforms, which links to Stpasha's comment about Heisenberg's uncertainty principle above). Clearly things are simpler to visualise if the characteristic function is real, which is the case if the pdf is symmetric about zero (that's in the 'Properties' section of this article). About the simplest such example to derive in probability theory is a Bernoulli distribution taking values –½ and +½ with probabilities p= ½ and 1 – p = ½. Then (similar to Stpasha's derivation immediately above) the characteristic function is ½e½it + ½e–½it = cos(½t). Another nice one to look at (not quite so easy to derive) is a uniform distribution (continuous) on [–½, ½], which in waveform terminology is a rectangular function (give or take a normalization constant) with characteristic function / Fourier transform sinc(t). I've never seen this in a probability textbook so it's kind of WP:original research, but if you think it helps i might put it in the article when i have time. Have to say, though, that as the main use of characteristic functions is for mathematical proofs, i'm not sure gaining intuition about their behaviour and relationship to pdfs is really much use. Perhaps it would help with their use in data analysis, but i've not used them for that myself. Qwfp (talk) 08:11, 26 May 2010 (UTC)
 * The specific characteristic function mentioned is "well-known" ... at least it is in the table of characteristic functions that I have just added as a reference. Melcombe (talk) 09:17, 26 May 2010 (UTC)
 * Sorry i finished writing previous comment in a hurry earlier, resulting in an "it" with an unintended referent as not everything in my head got as far as the keyboard. I was envisaging pairs of graphs of a few symmetric pdfs and their characteristic functions as i wrote it, and it's the use of such graphs to illustrate characteristic functions that i haven't seen in a textbook. On reflection i don't think anyone could claim that would be original research though. That sinc(t) is the characteristic function for U(–½, ½) follows immediately from the general case of the uniform already given in the Examples section. Qwfp (talk) 11:56, 26 May 2010 (UTC)

How about explaining what "E" is ? I think we shouldn't assume any universal meaning of any notation, especially when people come from various countries to this article and may not understand what "E" or "M" is. For example first sentence of introduction should be:

The characteristic function ( FX(x)) provides an alternative way for describing a random variable (X). Similarly to the cumulative distribution function and then explanation to what E in this formula is. This way every symbol used is explained in the description. I am new to wikipedia and I am not comfortable editing such important article yet. — Preceding unsigned comment added by Zenmaster82 (talk • contribs) 21:10, 18 November 2011 (UTC)

Definition for k×p-dimensional random matrix
What's T suppose to denote? How different is this from t? —Preceding unsigned comment added by 161.53.64.70 (talk) 08:53, 7 October 2009 (UTC)


 * It was supposed to be small t — the k×p matrix argument of the c.f.  …  st pasha  »  19:58, 7 October 2009 (UTC)

"and the last formula in parentheses is valid"
This text appears in the Definition Section. .. I may be blind, but I don't see any parentheses in that section.. OldMacDonalds (talk) 03:45, 19 February 2014 (UTC)

Why only for strictly positive?
"Inversion formulas", after the Levy theorem: "This formula is valid only for strictly positive random variables." — Really?? Hard to believe. What exactly is meant? (This is a recent insertion by User:Manoguru.) Boris Tsirelson (talk) 20:07, 29 July 2015 (UTC)
 * Says so in N.G. Shepard's (1991a), "A Framework for Inversion", pg. 521 in the paragraph discussing Levy's formula. He cites this as the reason for the need for Gil-Peleaz formula. Manoguru (talk) 23:30, 29 July 2015 (UTC)
 * Rather strange. The link to Shephard 1991a does not work (for me today, at least). And here is a quote from a textbook :
 * The inversion formula. Let $$\varphi(t) = \int e^{itx} \, \mu(dx)$$ where $$\mu$$ is a probability measure. If $$a<b$$ then
 * $$\displaystyle \lim_{T\to\infty} (2\pi)^{-1} \int_{-T}^T \frac{e^{-ita}-e^{-itb}}{it} \varphi(t) \, dt = \mu(a,b) + \frac12 \mu(\{a,b\}).$$
 * Boris Tsirelson (talk) 20:30, 30 July 2015 (UTC)


 * Well, I see what happens, after looking at the articles by Shepard and Gil-Pelaez. They do not doubt that Levy theorem holds for all random variables. The point is quite different. Levy theorem gives $$F(b)-F(a)$$ but not $$F(b)$$ alone. Theoretically, this is not a problem, since $$F(b)=\lim_{a\to-\infty} (F(b)-F(a)).$$ But numerically, it is rather impractical to send $$a$$ to $$-\infty.$$
 * I correct the article accordingly.
 * Boris Tsirelson (talk) 21:04, 30 July 2015 (UTC)


 * Thanks for the reply. I don't have access to Gil-Peleaz's original paper, which appears in Biometrika and our university does not have subscription to it, sadly. So I could not verify exactly why or under what condition this is true, if at all. Also, there is a bit of a confusion in the way the comment is stated. I was not sure if the restriction applies to the Levy's formula or to the re-statement that is amenable for numerical computation. Nevertheless, I felt it important to put it in the record. After all, it does get you to think why Gil-Peleaz is necessary if Levy is fine enough by itself. Best Regards, Manoguru (talk) 21:14, 30 July 2015 (UTC) P.S. Thanks for the reference by Durrett. I found a free ebook, and its fantastic. :)
 * Yes, I like that book, and cite it in "Conditioning (probability)" and other articles.
 * On the other hand, you could present your findings here on the talk page, given that you did not have enough information. Boris Tsirelson (talk) 21:24, 30 July 2015 (UTC)
 * I will try to keep that in mind. Manoguru (talk) 21:45, 30 July 2015 (UTC)

Relation to independence/ Kac's Theorem
This [1] looks like a very useful statement, relating independence of $$X,Y$$ to the characteristic functions $$\varphi_{(X,Y)}$$ and $$\varphi_{X},\varphi_{Y}$$. Is there a reason for the absence? (If not, I can add it.)

Theorem (Kac's theorem, from [1]). Let $$X,Y \in L^1$$ be $$\mathbb{R}^d$$-valued random variables. Then the following statements are equivalent. 1. $$X,Y$$ are independent 2. $$\forall \eta,\xi \in \mathbb{R}^d$$, $$\mathbb{E}e^{i \langle(X,Y), (\xi,\eta)\rangle} = \mathbb{E}e^{i \, X \cdot \xi} \cdot \mathbb{E}e^{i \, Y \cdot \eta}$$

[1] http://math.stackexchange.com/questions/287138/moment-generating-functions-characteristic-functions-of-x-y-factor-implies-x/287321#287321 — Preceding unsigned comment added by Ceacy (talk • contribs) 11:01, 30 May 2016 (UTC)

This can't be right: F(A) - F(A - 0)?
Theorem. If a is (possibly) an atom of X (in the univariate case this means a point of discontinuity of FX ) then
 * If X is scalar:
 * $$F_X(a) - F_X(a-0) = \lim_{T\to\infty}\frac{1}{2T}\int_{-T}^{+T}e^{-ita}\varphi_X(t)\,dt$$

But $$F_X(a-0) = F_X(a)$$ so this quantity is trivially zero. Is this supposed to reflect some kind of limit?

Manybytes (talk) 03:48, 24 September 2019 (UTC)


 * Sure, this is (a usual, convenient notation for) a limit: $$F_X(a-0) = \lim_{b\to a,b<a} F_X(b).$$ See one-sided limit. Well, $$F_X(a-)$$ is better; I guess, $$F_X(a-0)$$ is out of fashion. Boris Tsirelson (talk) 04:25, 24 September 2019 (UTC)

Is it really the Fourier transform, or only *almost* the Fourier transform?
The introduction contains this sentence:

"If a random variable admits a probability density function, then the characteristic function is the Fourier transform of the probability density function."

But among the various definitions of the Fourier transform, all of them involve an integral involving exp(-ixt) or exp(-2πixt), with a negative sign inside the exponential.

Doesn't this mean that the characteristic function is more like the inverse Fourier transform of the density function (if any)? 2601:200:C000:1A0:BC00:5039:DB55:E9EC (talk) 18:51, 29 July 2022 (UTC)


 * I agree that the inverse Fourier transform is closer than the forward Fourier transform. But there is also the factor of 2$\pi$ to worry about.  The inverse Fourier transform is something like $$\int F(\nu) e^{2 \pi \nu t} d\nu$$.  With the substitution $$\omega = 2 \pi \nu$$ that becomes $$\frac{1}{2\pi} \int F(\omega/2\pi) e^{\omega t} d\omega$$.  Either way, there is a factor of 2π to be reckoned with.  What if we say that the characteristic function is closely related to the inverse Fourier transform?  — Q uantling (talk &#124; contribs) 19:02, 29 July 2022 (UTC)