## Two Types of Random Variables

A random variable [latex]text{x}[/latex], and its distribution, can be discrete or continuous.

### Learning Objectives

Contrast discrete and continuous variables

### Key Takeaways

#### Key Points

- A random variable is a variable taking on numerical values determined by the outcome of a random phenomenon.
- The probability distribution of a random variable [latex]text{x}[/latex] tells us what the possible values of [latex]text{x}[/latex] are and what probabilities are assigned to those values.
- A discrete random variable has a countable number of possible values.
- The probability of each value of a discrete random variable is between 0 and 1, and the sum of all the probabilities is equal to 1.
- A continuous random variable takes on all the values in some interval of numbers.
- A density curve describes the probability distribution of a continuous random variable, and the probability of a range of events is found by taking the area under the curve.

#### Key Terms

**random variable**: a quantity whose value is random and to which a probability distribution is assigned, such as the possible outcome of a roll of a die**discrete random variable**: obtained by counting values for which there are no in-between values, such as the integers 0, 1, 2, ….**continuous random variable**: obtained from data that can take infinitely many values

### Random Variables

In probability and statistics, a randomvariable is a variable whose value is subject to variations due to chance (i.e. randomness, in a mathematical sense). As opposed to other mathematical variables, a random variable conceptually does not have a single, fixed value (even if unknown); rather, it can take on a set of possible different values, each with an associated probability.

A random variable’s possible values might represent the possible outcomes of a yet-to-be-performed experiment, or the possible outcomes of a past experiment whose already-existing value is uncertain (for example, as a result of incomplete information or imprecise measurements). They may also conceptually represent either the results of an “objectively” random process (such as rolling a die), or the “subjective” randomness that results from incomplete knowledge of a quantity.

Random variables can be classified as either discrete (that is, taking any of a specified list of exact values) or as continuous (taking any numerical value in an interval or collection of intervals). The mathematical function describing the possible values of a random variable and their associated probabilities is known as a probability distribution.

### Discrete Random Variables

Discrete random variables can take on either a finite or at most a countably infinite set of discrete values (for example, the integers). Their probability distribution is given by a probability mass function which directly maps each value of the random variable to a probability. For example, the value of [latex]text{x}_1[/latex] takes on the probability [latex]text{p}_1[/latex], the value of [latex]text{x}_2[/latex] takes on the probability [latex]text{p}_2[/latex], and so on. The probabilities [latex]text{p}_text{i}[/latex] must satisfy two requirements: every probability [latex]text{p}_text{i}[/latex] is a number between 0 and 1, and the sum of all the probabilities is 1. ([latex]text{p}_1+text{p}_2+dots + text{p}_text{k} = 1[/latex])

Examples of discrete random variables include the values obtained from rolling a die and the grades received on a test out of 100.

### Continuous Random Variables

Continuous random variables, on the other hand, take on values that vary continuously within one or more real intervals, and have a cumulative distribution function (CDF) that is absolutely continuous. As a result, the random variable has an uncountable infinite number of possible values, all of which have probability 0, though ranges of such values can have nonzero probability. The resulting probability distribution of the random variable can be described by a probability density, where the probability is found by taking the area under the curve.

Selecting random numbers between 0 and 1 are examples of continuous random variables because there are an infinite number of possibilities.

## Probability Distributions for Discrete Random Variables

Probability distributions for discrete random variables can be displayed as a formula, in a table, or in a graph.

### Learning Objectives

Give examples of discrete random variables

### Key Takeaways

#### Key Points

- A discrete probability function must satisfy the following: [latex]0 leq text{f}(text{x}) leq 1[/latex], i.e., the values of [latex]text{f}(text{x})[/latex] are probabilities, hence between 0 and 1.
- A discrete probability function must also satisfy the following: [latex]sum text{f}(text{x}) = 1[/latex], i.e., adding the probabilities of all disjoint cases, we obtain the probability of the sample space, 1.
- The probability mass function has the same purpose as the probability histogram, and displays specific probabilities for each discrete random variable. The only difference is how it looks graphically.

#### Key Terms

**discrete random variable**: obtained by counting values for which there are no in-between values, such as the integers 0, 1, 2, ….**probability distribution**: A function of a discrete random variable yielding the probability that the variable will have a given value.**probability mass function**: a function that gives the relative probability that a discrete random variable is exactly equal to some value

A discrete random variable [latex]text{x}[/latex] has a countable number of possible values. The probability distribution of a discrete random variable [latex]text{x}[/latex] lists the values and their probabilities, where value [latex]text{x}_1[/latex] has probability [latex]text{p}_1[/latex], value [latex]text{x}_2[/latex] has probability [latex]text{x}_2[/latex], and so on. Every probability [latex]text{p}_text{i}[/latex] is a number between 0 and 1, and the sum of all the probabilities is equal to 1.

Examples of discrete random variables include:

- The number of eggs that a hen lays in a given day (it can’t be 2.3)
- The number of people going to a given soccer match
- The number of students that come to class on a given day
- The number of people in line at McDonald’s on a given day and time

A discrete probability distribution can be described by a table, by a formula, or by a graph. For example, suppose that [latex]text{x}[/latex] is a random variable that represents the number of people waiting at the line at a fast-food restaurant and it happens to only take the values 2, 3, or 5 with probabilities [latex]frac{2}{10}[/latex], [latex]frac{3}{10}[/latex], and [latex]frac{5}{10}[/latex] respectively. This can be expressed through the function [latex]text{f}(text{x})= frac{text{x}}{10}[/latex]*, *[latex]text{x}=2, 3, 5[/latex] or through the table below. Of the conditional probabilities of the event [latex]text{B}[/latex] given that [latex]text{A}_1[/latex] is the case or that [latex]text{A}_2[/latex] is the case, respectively. Notice that these two representations are equivalent, and that this can be represented graphically as in the probability histogram below.

The formula, table, and probability histogram satisfy the following necessary conditions of discrete probability distributions:

- [latex]0 leq text{f}(text{x}) leq 1[/latex], i.e., the values of [latex]text{f}(text{x})[/latex] are probabilities, hence between 0 and 1.
- [latex]sum text{f}(text{x}) = 1[/latex], i.e., adding the probabilities of all disjoint cases, we obtain the probability of the sample space, 1.

Sometimes, the discrete probability distribution is referred to as the probability mass function (pmf). The probability mass function has the same purpose as the probability histogram, and displays specific probabilities for each discrete random variable. The only difference is how it looks graphically.

## Expected Values of Discrete Random Variables

The expected value of a random variable is the weighted average of all possible values that this random variable can take on.

### Learning Objectives

Calculate the expected value of a discrete random variable

### Key Takeaways

#### Key Points

- The expected value of a random variable [latex]text{X}[/latex] is defined as: [latex]text{E}[text{X}] = text{x}_1text{p}_1 + text{x}_2text{p}_2 + dots + text{x}_text{i}text{p}_text{i}[/latex], which can also be written as: [latex]text{E}[text{X}] = sum text{x}_text{i}text{p}_text{i}[/latex].
- If all outcomes [latex]text{x}_text{i}[/latex] are equally likely (that is, [latex]text{p}_1=text{p}_2=dots = text{p}_text{i}[/latex]), then the weighted average turns into the simple average.
- The expected value of [latex]text{X}[/latex] is what one expects to happen on average, even though sometimes it results in a number that is impossible (such as 2.5 children).

#### Key Terms

**discrete random variable**: obtained by counting values for which there are no in-between values, such as the integers 0, 1, 2, ….**expected value**: of a discrete random variable, the sum of the probability of each possible outcome of the experiment multiplied by the value itself

### Discrete Random Variable

A discrete random variable [latex]text{X}[/latex] has a countable number of possible values. The probability distribution of a discrete random variable [latex]text{X}[/latex] lists the values and their probabilities, such that [latex]text{x}_text{i}[/latex] has a probability of [latex]text{p}_text{i}[/latex]. The probabilities [latex]text{p}_text{i}[/latex] must satisfy two requirements:

- Every probability [latex]text{p}_text{i}[/latex] is a number between 0 and 1.
- The sum of the probabilities is 1: [latex]text{p}_1+text{p}_2+dots + text{p}_text{i} = 1[/latex].

### Expected Value Definition

In probability theory, the expected value (or expectation, mathematical expectation, EV, mean, or first moment) of a random variable is the weighted average of all possible values that this random variable can take on. The weights used in computing this average are probabilities in the case of a discrete random variable.

The expected value may be intuitively understood by the law of large numbers: the expected value, when it exists, is almost surely the limit of the sample mean as sample size grows to infinity. More informally, it can be interpreted as the long-run average of the results of many independent repetitions of an experiment (e.g. a dice roll). The value may not be expected in the ordinary sense—the “expected value” itself may be unlikely or even impossible (such as having 2.5 children), as is also the case with the sample mean.

### How To Calculate Expected Value

Suppose random variable [latex]text{X}[/latex] can take value [latex]text{x}_1[/latex] with probability [latex]text{p}_1[/latex], value [latex]text{x}_2[/latex] with probability [latex]text{p}_2[/latex], and so on, up to value [latex]text{x}_i[/latex] with probability [latex]text{p}_i[/latex]. Then the expectation value of a random variable [latex]text{X}[/latex] is defined as: [latex]text{E}[text{X}] = text{x}_1text{p}_1 + text{x}_2text{p}_2 + dots + text{x}_text{i}text{p}_text{i}[/latex], which can also be written as: [latex]text{E}[text{X}] = sum text{x}_text{i}text{p}_text{i}[/latex].

If all outcomes [latex]text{x}_text{i}[/latex] are equally likely (that is, [latex]text{p}_1 = text{p}_2 = dots = text{p}_text{i}[/latex]), then the weighted average turns into the simple average. This is intuitive: the expected value of a random variable is the average of all values it can take; thus the expected value is what one expects to happen on average. If the outcomes [latex]text{x}_text{i}[/latex] are not equally probable, then the simple average must be replaced with the weighted average, which takes into account the fact that some outcomes are more likely than the others. The intuition, however, remains the same: the expected value of [latex]text{X}[/latex] is what one expects to happen on average.

For example, let [latex]text{X}[/latex] represent the outcome of a roll of a six-sided die. The possible values for [latex]text{X}[/latex] are 1, 2, 3, 4, 5, and 6, all equally likely (each having the probability of [latex]frac{1}{6}[/latex]). The expectation of [latex]text{X}[/latex] is: [latex]text{E}[text{X}] = frac{1text{x}_1}{6} + frac{2text{x}_2}{6} + frac{3text{x}_3}{6} + frac{4text{x}_4}{6} + frac{5text{x}_5}{6} + frac{6text{x}_6}{6} = 3.5[/latex]. In this case, since all outcomes are equally likely, we could have simply averaged the numbers together: [latex]frac{1+2+3+4+5+6}{6} = 3.5[/latex].

Source: Statistics