Variance of sine and cosine of a random variableCalculate the conditional variance of exponential...
Why does this quiz question say that protons and electrons do not combine to form neutrons?
How can a kingdom keep the secret of a missing monarch from the public?
Aligning Systems of Equations
Multiplying elements of a list
How do I split ammo?
Boss asked me to sign a resignation paper without a date on it along with my new contract
Have any astronauts or cosmonauts died in space?
Can I legally make a website about boycotting a certain company?
What is formjacking?
Did the characters in Moving Pictures not know about cameras like Twoflower's?
How do I avoid the "chosen hero" feeling?
Why write a book when there's a movie in my head?
Is it possible to detect 100% of SQLi with a simple regex?
How do you pronounce "Aeclepiadae"?
What did Putin say about a US deep state in his state-of-the-nation speech; what has he said in the past?
Build ASCII Podiums
Why are "square law" devices important?
What if you do not believe in the project benefits?
Are all power cords made equal?
Why don't you get burned by the wood benches in a sauna?
Coworker asking me to not bring cakes due to self control issue. What should I do?
Why Doesn't It Completely Uninstall?
Can a planet be tidally unlocked?
Coworker is trying to get me to sign his petition to run for office. How to decline politely?
Variance of sine and cosine of a random variable
Calculate the conditional variance of exponential distribution with a constant value shift of the random variableWhat does “variance unity” mean?: “A normal distribution with mean $mu$ and variance unity”.Why square a constant when determining variance of a random variable?Variance of the Sample Variance of a normal distributionMean and variance of a Random variableVariance(XY + X/Y)Finding the mean and variance of random variables (discrete and continuous, specifically of indicators) and their properties.compute confidence interval from “variance of the sample variance”?Maximum likelihood estimator of the difference between two normal means and minimising its variancegenerating sample from binomial distribution and calculate the mean and variance.
$begingroup$
Suppose $X$ is a random variable drawn from normal distribution with mean $E$ and variance $V$. How could I calculate variance of $sin(X)$ and $cos(X)$?
(I thought the question was simple and tried to do a search, but did not find any good answer.)
What if there is no assumption about the distribution of $X$, and only sample mean and variance are provided?
variance
New contributor
$endgroup$
add a comment |
$begingroup$
Suppose $X$ is a random variable drawn from normal distribution with mean $E$ and variance $V$. How could I calculate variance of $sin(X)$ and $cos(X)$?
(I thought the question was simple and tried to do a search, but did not find any good answer.)
What if there is no assumption about the distribution of $X$, and only sample mean and variance are provided?
variance
New contributor
$endgroup$
1
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago
add a comment |
$begingroup$
Suppose $X$ is a random variable drawn from normal distribution with mean $E$ and variance $V$. How could I calculate variance of $sin(X)$ and $cos(X)$?
(I thought the question was simple and tried to do a search, but did not find any good answer.)
What if there is no assumption about the distribution of $X$, and only sample mean and variance are provided?
variance
New contributor
$endgroup$
Suppose $X$ is a random variable drawn from normal distribution with mean $E$ and variance $V$. How could I calculate variance of $sin(X)$ and $cos(X)$?
(I thought the question was simple and tried to do a search, but did not find any good answer.)
What if there is no assumption about the distribution of $X$, and only sample mean and variance are provided?
variance
variance
New contributor
New contributor
edited 2 hours ago
Hùng Phạm
New contributor
asked 2 hours ago
Hùng PhạmHùng Phạm
262
262
New contributor
New contributor
1
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago
add a comment |
1
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago
1
1
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
What si below is for $mu=0$ (and variance renamed $sigma^2$). Then $mathbb{E}[sin X]=0$, and you have
$$
operatorname{Var} sin X = mathbb{E}[sin^2 X]
= frac{1}{2}left(1-mathbb{E}[cos 2X]right)
$$
and
$$
mathbb{E}[cos 2X] = sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} mathbb{E}[X^{2k}]
= sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} sigma^{2k} (2k-1)!!
= sum_{n=0}^infty (-1)^k frac{2^{k}sigma^{2k}}{k!} = e^{-2sigma^{2}}
$$
and therefore
$$
operatorname{Var} sin X = boxed{frac{1-e^{-2sigma^2}}{2}}
$$
You can deal with the variance of $cos X$ in a similar fashion (but you now have to substract a non-zero $mathbb{E}[cos X]^2$), especially recalling that $mathbb{E}[cos^2 X] = 1- mathbb{E}[sin^2 X]$.
Now, for non-zero mean $mu$, you have
$$
sin(X-mu) = sin Xcos mu - cos Xsinmu
$$
(and similarly for $cos(X-mu)$)
Since $X-mu$ is a zero-mean Gaussian with variance $sigma^2$, we have computed the mean and variance of $sin(X-mu)$, $cos(X-mu)$ already. You can use this with the above trigonometric identities to find those of $cos X$ and $sin X$. (it's a bit cumbersome, but not too hard.)
Without knowing anything about the distribution of $X$, I don't think there's much you can do.
$endgroup$
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
|
show 4 more comments
$begingroup$
Here is a general formulation using the law of the unconscious statistician. that can be applied to other functions too. For specific calculations with $sin$ and $cos$ here though, I would say Clement C.'s answer is better!
The mean of $color{blue}{h(X)}$ (for some function $h$) would be given by the integral
$$mathbb{E}[h(X)]=int_{-infty}^{infty}color{blue}{h(x)}f_X(x), dx,$$
where $f_X$ is the probability density function of $X$.
The second moment would be found similarly as $$mathbb{E}left[(h(X))^2right] = int_{-infty}^{infty}color{blue}{(h(x)^2)}f_X(x), dx.$$
Once you know the first two moments here, you can calculate the variance using $mathrm{Var}(Z) = mathbb{E}[Z^2] - (mathbb{E}[Z])^2$.
Replace $h(x)$ with $cos x$ for the corresponding expectations for $cos X$, and similarly with $sin x$.
If the distribution of $X$ is not known, we cannot generally compute the exact mean and variance of $h(X)$. However, you may want to see this for some approximations that could be used. Some useful ones for you may be that if $X$ has mean $mu_X$ and variance $sigma^2_X$, then
$$mathbb{E}[h(X)]approx h(mu_X) + dfrac{h''(mu_X)}{2}sigma_X^2$$
and
$$mathrm{Var}(h(X))approx (h'(mu_X)^2)sigma^2_X + dfrac{1}{2}(h''(mu_X))^2 sigma^4_X.$$
$endgroup$
add a comment |
$begingroup$
$cos^2(x) = frac{cos(2x)+1}2$, which averages out to $frac12$. So as the variance of $X$ goes to infinity, the variance of $cos(X)$ goes to $frac12$, assuming the distribution of $X$ is "well-behaved". The lower bound is $0$ (the variance can be made arbitrarily small by choosing the variance of $X$ to be small enough), and as @angryavian says, the upper bound is $1$. Since $|cos(x)| leq 0$, and the inequality is strict for all but a measure zero set, the variance of $cos(X)$ is less than the variance of $X$.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Hùng Phạm is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3121914%2fvariance-of-sine-and-cosine-of-a-random-variable%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
What si below is for $mu=0$ (and variance renamed $sigma^2$). Then $mathbb{E}[sin X]=0$, and you have
$$
operatorname{Var} sin X = mathbb{E}[sin^2 X]
= frac{1}{2}left(1-mathbb{E}[cos 2X]right)
$$
and
$$
mathbb{E}[cos 2X] = sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} mathbb{E}[X^{2k}]
= sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} sigma^{2k} (2k-1)!!
= sum_{n=0}^infty (-1)^k frac{2^{k}sigma^{2k}}{k!} = e^{-2sigma^{2}}
$$
and therefore
$$
operatorname{Var} sin X = boxed{frac{1-e^{-2sigma^2}}{2}}
$$
You can deal with the variance of $cos X$ in a similar fashion (but you now have to substract a non-zero $mathbb{E}[cos X]^2$), especially recalling that $mathbb{E}[cos^2 X] = 1- mathbb{E}[sin^2 X]$.
Now, for non-zero mean $mu$, you have
$$
sin(X-mu) = sin Xcos mu - cos Xsinmu
$$
(and similarly for $cos(X-mu)$)
Since $X-mu$ is a zero-mean Gaussian with variance $sigma^2$, we have computed the mean and variance of $sin(X-mu)$, $cos(X-mu)$ already. You can use this with the above trigonometric identities to find those of $cos X$ and $sin X$. (it's a bit cumbersome, but not too hard.)
Without knowing anything about the distribution of $X$, I don't think there's much you can do.
$endgroup$
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
|
show 4 more comments
$begingroup$
What si below is for $mu=0$ (and variance renamed $sigma^2$). Then $mathbb{E}[sin X]=0$, and you have
$$
operatorname{Var} sin X = mathbb{E}[sin^2 X]
= frac{1}{2}left(1-mathbb{E}[cos 2X]right)
$$
and
$$
mathbb{E}[cos 2X] = sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} mathbb{E}[X^{2k}]
= sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} sigma^{2k} (2k-1)!!
= sum_{n=0}^infty (-1)^k frac{2^{k}sigma^{2k}}{k!} = e^{-2sigma^{2}}
$$
and therefore
$$
operatorname{Var} sin X = boxed{frac{1-e^{-2sigma^2}}{2}}
$$
You can deal with the variance of $cos X$ in a similar fashion (but you now have to substract a non-zero $mathbb{E}[cos X]^2$), especially recalling that $mathbb{E}[cos^2 X] = 1- mathbb{E}[sin^2 X]$.
Now, for non-zero mean $mu$, you have
$$
sin(X-mu) = sin Xcos mu - cos Xsinmu
$$
(and similarly for $cos(X-mu)$)
Since $X-mu$ is a zero-mean Gaussian with variance $sigma^2$, we have computed the mean and variance of $sin(X-mu)$, $cos(X-mu)$ already. You can use this with the above trigonometric identities to find those of $cos X$ and $sin X$. (it's a bit cumbersome, but not too hard.)
Without knowing anything about the distribution of $X$, I don't think there's much you can do.
$endgroup$
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
|
show 4 more comments
$begingroup$
What si below is for $mu=0$ (and variance renamed $sigma^2$). Then $mathbb{E}[sin X]=0$, and you have
$$
operatorname{Var} sin X = mathbb{E}[sin^2 X]
= frac{1}{2}left(1-mathbb{E}[cos 2X]right)
$$
and
$$
mathbb{E}[cos 2X] = sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} mathbb{E}[X^{2k}]
= sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} sigma^{2k} (2k-1)!!
= sum_{n=0}^infty (-1)^k frac{2^{k}sigma^{2k}}{k!} = e^{-2sigma^{2}}
$$
and therefore
$$
operatorname{Var} sin X = boxed{frac{1-e^{-2sigma^2}}{2}}
$$
You can deal with the variance of $cos X$ in a similar fashion (but you now have to substract a non-zero $mathbb{E}[cos X]^2$), especially recalling that $mathbb{E}[cos^2 X] = 1- mathbb{E}[sin^2 X]$.
Now, for non-zero mean $mu$, you have
$$
sin(X-mu) = sin Xcos mu - cos Xsinmu
$$
(and similarly for $cos(X-mu)$)
Since $X-mu$ is a zero-mean Gaussian with variance $sigma^2$, we have computed the mean and variance of $sin(X-mu)$, $cos(X-mu)$ already. You can use this with the above trigonometric identities to find those of $cos X$ and $sin X$. (it's a bit cumbersome, but not too hard.)
Without knowing anything about the distribution of $X$, I don't think there's much you can do.
$endgroup$
What si below is for $mu=0$ (and variance renamed $sigma^2$). Then $mathbb{E}[sin X]=0$, and you have
$$
operatorname{Var} sin X = mathbb{E}[sin^2 X]
= frac{1}{2}left(1-mathbb{E}[cos 2X]right)
$$
and
$$
mathbb{E}[cos 2X] = sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} mathbb{E}[X^{2k}]
= sum_{n=0}^infty (-1)^kfrac{2^{2k}}{(2k)!} sigma^{2k} (2k-1)!!
= sum_{n=0}^infty (-1)^k frac{2^{k}sigma^{2k}}{k!} = e^{-2sigma^{2}}
$$
and therefore
$$
operatorname{Var} sin X = boxed{frac{1-e^{-2sigma^2}}{2}}
$$
You can deal with the variance of $cos X$ in a similar fashion (but you now have to substract a non-zero $mathbb{E}[cos X]^2$), especially recalling that $mathbb{E}[cos^2 X] = 1- mathbb{E}[sin^2 X]$.
Now, for non-zero mean $mu$, you have
$$
sin(X-mu) = sin Xcos mu - cos Xsinmu
$$
(and similarly for $cos(X-mu)$)
Since $X-mu$ is a zero-mean Gaussian with variance $sigma^2$, we have computed the mean and variance of $sin(X-mu)$, $cos(X-mu)$ already. You can use this with the above trigonometric identities to find those of $cos X$ and $sin X$. (it's a bit cumbersome, but not too hard.)
Without knowing anything about the distribution of $X$, I don't think there's much you can do.
edited 1 hour ago
answered 2 hours ago
Clement C.Clement C.
50.5k33891
50.5k33891
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
|
show 4 more comments
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
I might have missed something, but could you tell me why $Var sin(X) = E[sin^2 X]$ ? Also why does the result not contain anything related to mean or variance of $X$?
$endgroup$
– Hùng Phạm
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm Oh, my bad, I did it for mean $0$ and variance $1$. Let me fix that.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
I think you'd need to consider $mathbb E(sin^2x)-[mathbb E(sinx)]^2$
$endgroup$
– Sharat V Chandrasekhar
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@SharatVChandrasekhar It becomes a bit more complicated than that, actually.
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
$begingroup$
@HùngPhạm I added the variance. Trying to see if this approach is amenable to a non-zero mean as well without too much pain..
$endgroup$
– Clement C.
2 hours ago
|
show 4 more comments
$begingroup$
Here is a general formulation using the law of the unconscious statistician. that can be applied to other functions too. For specific calculations with $sin$ and $cos$ here though, I would say Clement C.'s answer is better!
The mean of $color{blue}{h(X)}$ (for some function $h$) would be given by the integral
$$mathbb{E}[h(X)]=int_{-infty}^{infty}color{blue}{h(x)}f_X(x), dx,$$
where $f_X$ is the probability density function of $X$.
The second moment would be found similarly as $$mathbb{E}left[(h(X))^2right] = int_{-infty}^{infty}color{blue}{(h(x)^2)}f_X(x), dx.$$
Once you know the first two moments here, you can calculate the variance using $mathrm{Var}(Z) = mathbb{E}[Z^2] - (mathbb{E}[Z])^2$.
Replace $h(x)$ with $cos x$ for the corresponding expectations for $cos X$, and similarly with $sin x$.
If the distribution of $X$ is not known, we cannot generally compute the exact mean and variance of $h(X)$. However, you may want to see this for some approximations that could be used. Some useful ones for you may be that if $X$ has mean $mu_X$ and variance $sigma^2_X$, then
$$mathbb{E}[h(X)]approx h(mu_X) + dfrac{h''(mu_X)}{2}sigma_X^2$$
and
$$mathrm{Var}(h(X))approx (h'(mu_X)^2)sigma^2_X + dfrac{1}{2}(h''(mu_X))^2 sigma^4_X.$$
$endgroup$
add a comment |
$begingroup$
Here is a general formulation using the law of the unconscious statistician. that can be applied to other functions too. For specific calculations with $sin$ and $cos$ here though, I would say Clement C.'s answer is better!
The mean of $color{blue}{h(X)}$ (for some function $h$) would be given by the integral
$$mathbb{E}[h(X)]=int_{-infty}^{infty}color{blue}{h(x)}f_X(x), dx,$$
where $f_X$ is the probability density function of $X$.
The second moment would be found similarly as $$mathbb{E}left[(h(X))^2right] = int_{-infty}^{infty}color{blue}{(h(x)^2)}f_X(x), dx.$$
Once you know the first two moments here, you can calculate the variance using $mathrm{Var}(Z) = mathbb{E}[Z^2] - (mathbb{E}[Z])^2$.
Replace $h(x)$ with $cos x$ for the corresponding expectations for $cos X$, and similarly with $sin x$.
If the distribution of $X$ is not known, we cannot generally compute the exact mean and variance of $h(X)$. However, you may want to see this for some approximations that could be used. Some useful ones for you may be that if $X$ has mean $mu_X$ and variance $sigma^2_X$, then
$$mathbb{E}[h(X)]approx h(mu_X) + dfrac{h''(mu_X)}{2}sigma_X^2$$
and
$$mathrm{Var}(h(X))approx (h'(mu_X)^2)sigma^2_X + dfrac{1}{2}(h''(mu_X))^2 sigma^4_X.$$
$endgroup$
add a comment |
$begingroup$
Here is a general formulation using the law of the unconscious statistician. that can be applied to other functions too. For specific calculations with $sin$ and $cos$ here though, I would say Clement C.'s answer is better!
The mean of $color{blue}{h(X)}$ (for some function $h$) would be given by the integral
$$mathbb{E}[h(X)]=int_{-infty}^{infty}color{blue}{h(x)}f_X(x), dx,$$
where $f_X$ is the probability density function of $X$.
The second moment would be found similarly as $$mathbb{E}left[(h(X))^2right] = int_{-infty}^{infty}color{blue}{(h(x)^2)}f_X(x), dx.$$
Once you know the first two moments here, you can calculate the variance using $mathrm{Var}(Z) = mathbb{E}[Z^2] - (mathbb{E}[Z])^2$.
Replace $h(x)$ with $cos x$ for the corresponding expectations for $cos X$, and similarly with $sin x$.
If the distribution of $X$ is not known, we cannot generally compute the exact mean and variance of $h(X)$. However, you may want to see this for some approximations that could be used. Some useful ones for you may be that if $X$ has mean $mu_X$ and variance $sigma^2_X$, then
$$mathbb{E}[h(X)]approx h(mu_X) + dfrac{h''(mu_X)}{2}sigma_X^2$$
and
$$mathrm{Var}(h(X))approx (h'(mu_X)^2)sigma^2_X + dfrac{1}{2}(h''(mu_X))^2 sigma^4_X.$$
$endgroup$
Here is a general formulation using the law of the unconscious statistician. that can be applied to other functions too. For specific calculations with $sin$ and $cos$ here though, I would say Clement C.'s answer is better!
The mean of $color{blue}{h(X)}$ (for some function $h$) would be given by the integral
$$mathbb{E}[h(X)]=int_{-infty}^{infty}color{blue}{h(x)}f_X(x), dx,$$
where $f_X$ is the probability density function of $X$.
The second moment would be found similarly as $$mathbb{E}left[(h(X))^2right] = int_{-infty}^{infty}color{blue}{(h(x)^2)}f_X(x), dx.$$
Once you know the first two moments here, you can calculate the variance using $mathrm{Var}(Z) = mathbb{E}[Z^2] - (mathbb{E}[Z])^2$.
Replace $h(x)$ with $cos x$ for the corresponding expectations for $cos X$, and similarly with $sin x$.
If the distribution of $X$ is not known, we cannot generally compute the exact mean and variance of $h(X)$. However, you may want to see this for some approximations that could be used. Some useful ones for you may be that if $X$ has mean $mu_X$ and variance $sigma^2_X$, then
$$mathbb{E}[h(X)]approx h(mu_X) + dfrac{h''(mu_X)}{2}sigma_X^2$$
and
$$mathrm{Var}(h(X))approx (h'(mu_X)^2)sigma^2_X + dfrac{1}{2}(h''(mu_X))^2 sigma^4_X.$$
edited 1 hour ago
answered 2 hours ago
Minus One-TwelfthMinus One-Twelfth
1,15819
1,15819
add a comment |
add a comment |
$begingroup$
$cos^2(x) = frac{cos(2x)+1}2$, which averages out to $frac12$. So as the variance of $X$ goes to infinity, the variance of $cos(X)$ goes to $frac12$, assuming the distribution of $X$ is "well-behaved". The lower bound is $0$ (the variance can be made arbitrarily small by choosing the variance of $X$ to be small enough), and as @angryavian says, the upper bound is $1$. Since $|cos(x)| leq 0$, and the inequality is strict for all but a measure zero set, the variance of $cos(X)$ is less than the variance of $X$.
$endgroup$
add a comment |
$begingroup$
$cos^2(x) = frac{cos(2x)+1}2$, which averages out to $frac12$. So as the variance of $X$ goes to infinity, the variance of $cos(X)$ goes to $frac12$, assuming the distribution of $X$ is "well-behaved". The lower bound is $0$ (the variance can be made arbitrarily small by choosing the variance of $X$ to be small enough), and as @angryavian says, the upper bound is $1$. Since $|cos(x)| leq 0$, and the inequality is strict for all but a measure zero set, the variance of $cos(X)$ is less than the variance of $X$.
$endgroup$
add a comment |
$begingroup$
$cos^2(x) = frac{cos(2x)+1}2$, which averages out to $frac12$. So as the variance of $X$ goes to infinity, the variance of $cos(X)$ goes to $frac12$, assuming the distribution of $X$ is "well-behaved". The lower bound is $0$ (the variance can be made arbitrarily small by choosing the variance of $X$ to be small enough), and as @angryavian says, the upper bound is $1$. Since $|cos(x)| leq 0$, and the inequality is strict for all but a measure zero set, the variance of $cos(X)$ is less than the variance of $X$.
$endgroup$
$cos^2(x) = frac{cos(2x)+1}2$, which averages out to $frac12$. So as the variance of $X$ goes to infinity, the variance of $cos(X)$ goes to $frac12$, assuming the distribution of $X$ is "well-behaved". The lower bound is $0$ (the variance can be made arbitrarily small by choosing the variance of $X$ to be small enough), and as @angryavian says, the upper bound is $1$. Since $|cos(x)| leq 0$, and the inequality is strict for all but a measure zero set, the variance of $cos(X)$ is less than the variance of $X$.
edited 1 hour ago
angryavian
41.5k23381
41.5k23381
answered 1 hour ago
AcccumulationAcccumulation
7,0192618
7,0192618
add a comment |
add a comment |
Hùng Phạm is a new contributor. Be nice, and check out our Code of Conduct.
Hùng Phạm is a new contributor. Be nice, and check out our Code of Conduct.
Hùng Phạm is a new contributor. Be nice, and check out our Code of Conduct.
Hùng Phạm is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3121914%2fvariance-of-sine-and-cosine-of-a-random-variable%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
Both variables are bounded in $[-1, 1]$ so one can show that the variance is $le frac{2^2}{4} = 1$.
$endgroup$
– angryavian
2 hours ago
$begingroup$
That's an interesting point, might be helpful so I'll keep it in mind. However, I would like to know an exact formula/procedure if it is possible.
$endgroup$
– Hùng Phạm
2 hours ago