Menu Top
Complete Course of Mathematics
Topic 1: Numbers & Numerical Applications Topic 2: Algebra Topic 3: Quantitative Aptitude
Topic 4: Geometry Topic 5: Construction Topic 6: Coordinate Geometry
Topic 7: Mensuration Topic 8: Trigonometry Topic 9: Sets, Relations & Functions
Topic 10: Calculus Topic 11: Mathematical Reasoning Topic 12: Vectors & Three-Dimensional Geometry
Topic 13: Linear Programming Topic 14: Index Numbers & Time-Based Data Topic 15: Financial Mathematics
Topic 16: Statistics & Probability


Content On This Page
Random Variables: Definition and Types (Discrete, Continuous) Probability Distribution: Definition and Properties (for Discrete Random Variables) Probability Distribution (Applied Maths Definition)


Random Variables and Probability Distributions




Random Variables: Definition and Types (Discrete, Continuous)


Definition of a Random Variable

In the study of probability, we often deal with outcomes of experiments that are not inherently numerical (e.g., heads or tails when tossing a coin, types of defects in manufacturing, colours of cars). However, to perform mathematical and statistical analysis on these outcomes, it is convenient to associate them with numerical values.

A **Random Variable** is a concept that provides a way to assign a specific numerical value to each possible outcome of a random experiment. It is not a variable in the algebraic sense (where it represents an unknown value), but rather a variable whose value is determined by the outcome of a random process.

Formally, a random variable, usually denoted by a capital letter like $X$, $Y$, or $Z$, is a **function** that maps each outcome in the sample space ($S$) of a random experiment to a unique real number.

$$X: S \to \mathbb{R}$$

... (1)

Where $S$ is the sample space of the random experiment, and $\mathbb{R}$ is the set of real numbers. For each outcome $s \in S$, the random variable $X$ assigns a specific numerical value $X(s)$. The set of all possible numerical values that the random variable can take is called its **range**.

Example: Experiment: Tossing two fair coins simultaneously.

The sample space is $S = \{HH, HT, TH, TT\}$.

Let the random variable $X$ be the "number of heads obtained" in this experiment.

The set of possible values (the range) that the random variable $X$ can take is $\{0, 1, 2\}$. The random variable transforms the non-numerical outcomes (HH, HT, etc.) into numerical outcomes (0, 1, 2).


Types of Random Variables

Random variables are categorized into two main types based on the set of values they can assume:

  1. Discrete Random Variable:

    A discrete random variable is a random variable whose possible values are countable. This means the variable can take on either a finite number of distinct values or a countably infinite number of distinct values.

    • The values of a discrete random variable often arise from **counting** processes.
    • There are distinct gaps between the possible values that the variable can take.

    Examples:

    • The number of heads in 3 coin tosses (possible values: 0, 1, 2, 3 - finite number).
    • The number of defective items in a sample of 10 items from a production line (possible values: 0, 1, 2, ..., 10 - finite number).
    • The sum of the numbers when rolling two dice (possible values: 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12 - finite number).
    • The number of phone calls received by a customer service center in an hour (possible values: 0, 1, 2, 3, ... - countably infinite number).
  2. Continuous Random Variable:

    A continuous random variable is a random variable that can take on any value within a given interval or range on the real number line. These values are infinite and uncountable.

    • The values of a continuous random variable typically arise from **measuring** processes.
    • Between any two possible values, there are infinitely many other possible values.

    Examples:

    • The height of a person (e.g., any value between 150 cm and 200 cm).
    • The temperature of a liquid (e.g., any value between 0°C and 100°C).
    • The time taken to complete a race (e.g., any value in seconds greater than 0).
    • The weight of a package (e.g., any value in kilograms greater than 0).

The distinction between discrete and continuous random variables is fundamental because the methods used to describe their probability distributions and calculate probabilities are different.



Probability Distribution: Definition and Properties (for Discrete Random Variables)


Definition

The **probability distribution** of a **discrete random variable** $X$ is a complete description of the probabilities associated with each possible value that the random variable can take. It tells us how the total probability of 1 is distributed among the different possible values of $X$.

A probability distribution can be presented in various forms, but it essentially provides a list or function of all possible values of $X$ and their corresponding probabilities.

Let the possible values that a discrete random variable $X$ can take be $x_1, x_2, x_3, \dots$, where the sequence of values may be finite or countably infinite. The probability distribution assigns a probability to each of these values, denoted by $P(X=x_i)$, which is often represented by $p_i$.


Representations of a Discrete Probability Distribution

A probability distribution for a discrete random variable can be presented in the following ways:

1. Table Form:

This is the most common way to represent the probability distribution for a finite number of possible values. The table lists each possible value of the random variable $X$ and its corresponding probability:

Value of X ($x_i$) Probability $P(X=x_i) = p_i$
$x_1$$p_1$
$x_2$$p_2$
$\vdots$$\vdots$
$x_n$$p_n$
Total$\sum p_i = 1$

Each $p_i$ must be calculated based on the underlying random experiment and the definition of the random variable.

2. Formula Form (Probability Mass Function - PMF):

For some discrete random variables, the probability $P(X=x)$ can be described by a mathematical formula, denoted as $p(x)$ or $f(x)$, which is called the **Probability Mass Function (PMF)**. The PMF gives the probability for each possible value $x$ of the random variable $X$.

$$p(x_i) = P(X=x_i) = p_i \quad \text{for each possible value } x_i$$

... (1)

3. Graphical Form:

A discrete probability distribution can be visualized using a bar chart or a line graph. The horizontal axis represents the possible values of the random variable ($x_i$), and the vertical axis represents the corresponding probabilities ($p_i$). The height of each bar (or the length of each line segment) indicates the probability of that specific value.

Bar chart representing a discrete probability distribution

Properties of a Discrete Probability Distribution

For a table, formula (PMF), or graph to be considered a valid probability distribution for a discrete random variable $X$, it must satisfy the following two fundamental properties, derived from the axioms of probability:

  1. Non-negativity:

    The probability of each possible value that the random variable can take must be greater than or equal to 0. Probabilities cannot be negative.

    $$P(X=x_i) = p_i \ge 0 \quad \text{for all possible values } x_i$$

    ... (2)

  2. Summation to One:

    The sum of the probabilities of all possible values that the random variable can take must be exactly equal to 1. This reflects that the random variable must take on one of its possible values.

    $$\sum P(X=x_i) = \sum p_i = 1$$

    ... (3)

    The summation is over all possible values of $X$. For a countably infinite set of values, this is an infinite series that must converge to 1.

Any set of probabilities that satisfies these two properties constitutes a valid discrete probability distribution.


Example

Example 1. Let X be the random variable representing the number of heads obtained when tossing two fair coins. Find the probability distribution of X.

Answer:

Given: Random variable X = number of heads in tossing two fair coins.

To Find: The probability distribution of X.

Solution:

The random experiment is tossing two fair coins. The sample space is $S = \{HH, HT, TH, TT\}$. The total number of equally likely outcomes is $n(S) = 4$. Since the coins are fair, the probability of each elementary outcome is $1/4$.

The random variable $X$ assigns a numerical value (number of heads) to each outcome:

  • $X(HH) = 2$
  • $X(HT) = 1$
  • $X(TH) = 1$
  • $X(TT) = 0$

The possible values that the random variable $X$ can take are $\{0, 1, 2\}$.

Now we calculate the probability for each possible value of $X$:

  • For $X=0$:

    This occurs for the outcome TT. $P(X=0) = P(\{TT\}) = \frac{1}{4}$.
  • For $X=1$:

    This occurs for the outcomes HT and TH. $P(X=1) = P(\{HT, TH\}) = P(\{HT\}) + P(\{TH\})$ (since outcomes are mutually exclusive) $= \frac{1}{4} + \frac{1}{4} = \frac{2}{4} = \frac{1}{2}$.
  • For $X=2$:

    This occurs for the outcome HH. $P(X=2) = P(\{HH\}) = \frac{1}{4}$.

The probability distribution of $X$ can be presented in a table format:

Value of X ($x_i$)
(No. of Heads)
Probability $P(X=x_i) = p_i$
01/4
11/2
21/4
Total$\frac{1}{4} + \frac{1}{2} + \frac{1}{4} = \frac{1+2+1}{4} = \frac{4}{4} = 1$

Check Properties:

1. All probabilities ($1/4, 1/2, 1/4$) are greater than or equal to 0. Property 1 is satisfied.

2. The sum of the probabilities is $1/4 + 1/2 + 1/4 = 1$. Property 2 is satisfied.

Since both properties are satisfied, this is a valid discrete probability distribution.




Probability Distribution (Applied Maths Definition)


From an applied mathematics or practical standpoint, a **probability distribution** is viewed as a **mathematical model** used to describe the behaviour of a random variable and quantify the likelihood of different possible outcomes for a random phenomenon observed in the real world. It goes beyond simply listing probabilities; it provides a structured way to represent and analyze uncertainty.


Key Aspects of Probability Distributions in Applied Contexts

In summary, from an applied perspective, probability distributions are not just abstract mathematical constructs but practical tools for modeling, analyzing, and making decisions in situations involving randomness and uncertainty, bridging the gap between theoretical probability and real-world data analysis.