Random Variables
Tags |
---|
What is a random variable?
A random variable is nothing more than a tuple containing distribution, and a range . This in indexes into , but that’s only an indexing property. So…as long as you have some set and a valid probability distribution with the same cardinality, you’ve got yourself a random variable.
Example
As an example, we can have be the and as the . Note how and can have no relation. As long as they have the same cardinality, it is possible to make an RV out of them.
If has more than one variable, it is a joint distribution
. When we express things like , we are now talking about a distribution because isn’t an event. Therefore, we can also write stuff like , which is a conditional probability distribution over two random variables. The chain rule and Bayes rule still applies
PMF, CDF, and PDF
- A PMF is , which is only a thing for discrete RV's
- A PDF is the PMF of continuous RV's, but this does NOT correspond to the probability\. Rather, it corresponds to the
density
of the distribution at that point.
- A CDF is , which is a summation for discrete RV's and an integral for continuous RV's.
The formal definition of density
So for continuous RV’s , we have the notion of density . When defining the density, we always do it in terms of the CDF .
so this is a great first step when you’re dealing with proofs with density. And of course, remember that
and you can apply functions inside the , like .
Same RV, same parameters
Here’s a critical distinction. When we talk about some RV , it has an identity as well as parameters. If we set , this means that every sample of is the same as .
However, there’s also the notion of sufficient statistics. If we have , then we have the same distribution, but the identity of are not the same. As a consequence, if you draw a sample from X, it may not be the same as the sample from Y.
Distributions over likelihoods
So is a random variable, and is also a random variable. This is because is just a function that takes in something and outputs a number between 0 and 1. Therefore, is feeding the sample back into its own likelihood fuction. This is perfectly valid, and in fact, this is exactly what entropy does! .