P(A)=Total number of possible outcomesNumber of favorable outcomes
Calculates the probability of event A happening when all outcomes are equally likely, quantifying how likely a specific event is to occur.
Probability of event A occurring.
Event whose probability is being calculated.
Count of outcomes that qualify as event A occurring.
Count of all possible outcomes in the sample space.
Flipping a fair coin, probability of heads: P(Heads)=21
Simple games of chance, basic statistical calculations.
Probability Range Of An Event
Formula:
0≤P(A)≤1
The probability of any event is quantified on a scale from 0 to 1, inclusive. A probability of 0 indicates an event is impossible and will not occur under any circumstances. A probability of 1 indicates an event is certain to occur. Probabilities between 0 and 1 indicate the likelihood of the event's occurrence, with values closer to 1 suggesting a higher probability and values closer to 0 suggesting a lower probability.
Event
Prabability of A.
0 - Rolling a 7 on a standard six-sided die; 1 - Drawing any card from a full deck of cards; Between 0 and 1 - Flipping a coin and it landing on heads, which has a probability of 0.5.
This fundamental concept is used in all fields involving probability calculations, including statistics, risk assessment, finance, science, and engineering.
Complement Rule
Formula:
P(A′)+P(A)=1
States that the probability of an event occurring plus the probability of its complement (the event not occurring) equals one. This is because the event and its complement together encompass all possible outcomes.
Probability of the complement of event A, which is the event not occurring.
Probability of event A occurring.
If there is a 30% chance it will rain tomorrow (P(A) = 0.30), then there is a 70% chance it will not rain (P(A') = 0.70).
Derived from the definition of probability space, where the total probability is 1. Thus, P(A′)=1−P(A).
Frequently used to simplify calculations where it is easier to calculate the chance of something not happening rather than it happening, such as in quality control or risk assessment scenarios.
Conditional Probability
Conditional Probability Basic Formula
Formula:
P(A∣B)=P(B)P(A∩B)
Measures the probability of event A occurring given that event B has already occurred, assessing how B influences A.
Probability of both events A and B happening together.
Probabilities of A and B where B's occurrence affects the likelihood of A.
Drawing two cards, probability second card is an ace given the first was an ace: P(Ace2∣Ace1)=513
Situations where prior information about one event affects the outcome of another.
Advanced Probabilities
Bayes' Theorem
Formula:
P(A∣B)=P(B)P(B∣A)×P(A)
Updates the probability of event A occurring based on the occurrence of event B, integrating prior belief (prior probability) with new evidence.
Initial or prior probability of A occurring, overall probability of B, and probability of B given A.
Events where B's occurrence provides evidence about A's likelihood.
Disease with 1% prevalence, 99% sensitive test, probability of disease given a positive test: Calculated using Bayes' Theorem.
Medical testing, machine learning algorithms, updating predictions with new information.
Independent Events
Probability Of Both Events Occurring (Multiplication Rule)
Formula:
P(A∩B)=P(A)×P(B)
Calculates the probability that both events A and B occur simultaneously. It assumes that A and B are independent, meaning the occurrence of A does not affect the occurrence of B, and vice versa.
Probability of both events A and B occurring together.
Probability of event A occurring independently, and probability of event B occurring independently.
Flipping two fair coins, the probability of both being heads: P(Heads∩Heads)=P(Heads)×P(Heads)=21×21=41
Used in scenarios involving multiple independent trials or events.
Probability Of Either Event Occurring (Addition Rule)
Formula:
P(A∪B)=P(A)+P(B)−P(A∩B)
This rule calculates the probability that at least one of the events A or B occurs, adjusting for the overlap by subtracting the probability of both occurring together.
Probability of either event A or event B occurring.
P(A∪B)=P(A)+P(B)−P(A)×P(B)
For rolling a die, the probability of rolling a 2 or a 5: P(2∪5)=P(2)+P(5)−P(2)×P(5)=61+61−61×61=3611
Useful in evaluating the likelihood of multiple possible outcomes where events are independent.
Probability Of At Least One Event Not Occurring
Formula:
P(¬A∩¬B)=P(¬A)×P(¬B)
Calculates the probability that neither event A nor event B occurs, using the complement rule for each event.
Probability of neither event A nor event B occurring.
Complements of probabilities of A and B occurring, respectively.
If the probability of rain on any given day is 0.20, then the probability of no rain for two independent days is: P(¬Rain1∩¬Rain2)=(1−0.20)×(1−0.20)=0.64
Important in scenarios where the non-occurrence of multiple independent events is critical, such as in reliability engineering.
Probability Of Exactly One Event Occurring
Formula:
P(exactly one of A or B)=P(A∩¬B)+P(¬A∩B)
Determines the probability of exactly one of the two events occurring, not both.
For independent events: P(exactly one of A or B)=P(A)×(1−P(B))+(1−P(A))×P(B)
Using a deck of cards, the probability that one draws an ace or a king, but not both: P(Ace∩¬King)+P(¬Ace∩King)
Used in analytical models where exclusive outcomes are analyzed.
General Formula For Multiple Independent Events
Formula:
P(A∩B∩C)=P(A)×P(B)×P(C)
Extends the multiplication rule to any number of events, calculating the probability that all specified independent events occur.
Probability of events A, B, and C all occurring together.
Individual Probabilities of A,B and C .
For three independent events each with a 50% chance of occurring, the probability of all occurring is: 0.5×0.5×0.5=0.125
Applicable in complex systems where multiple independent events must be simultaneously considered, such as in network reliability.
Disjoint Events
Probability Of Both Disjoint Events Occurring
Formula:
P(A∩B)=0
Since disjoint events cannot occur simultaneously, the probability of both occurring at the same time is zero. This formula pretty much holds the essense of disjoint events definition : chance that both of them happen is equal to zero.
Probability of both disjoint events A and B occurring together.
If event A is drawing a heart from a deck of cards and event B is drawing a spade, since one draw cannot be both a heart and a spade, P(A∩B)=0.
Applicable in any scenario where two or more outcomes are mutually exclusive.
Probability Of Either Disjoint Event Occurring (Addition Rule)
Formula:
P(A∪B)=P(A)+P(B)
The probability of either disjoint event occurring is the sum of their individual probabilities, because there is no overlap between them.
Probability of either disjoint event A or event B occurring.
If event A is drawing a heart (25% chance) and event B is drawing a spade (25% chance) from a deck, P(A∪B)=25.
Useful in determining the likelihood of one of several exclusive events occurring.
Probability Of Neither Disjoint Event Occurring
Formula:
P(¬A∩¬B)=P(¬A)×P(¬B)
Calculates the probability that neither disjoint event occurs, assuming each event’s non-occurrence is independent of the other.
Probability of neither disjoint event A nor event B occurring.
If the probability of not drawing a heart is 75% and not drawing a spade is 75%, then P(¬A∩¬B)=75.
Important for risk assessment and contingency planning in mutually exclusive scenarios.
Conditional Probability For Disjoint Events
Formula:
P(A∣B)=0 and P(B∣A)=0
If one disjoint event occurs, the other cannot possibly happen, thus the conditional probability is zero.
Probability of A given B and probability of B given A, respectively, for disjoint events.
If event A has already occurred (e.g., a heart was drawn), the probability of event B (drawing a spade) occurring is 0.
Used in scenarios where the occurrence of one outcome precludes the occurrence of others.
Generalization To Multiple Disjoint Events
Formula:
P(A∪B∪C∪…)=P(A)+P(B)+P(C)+…
For multiple disjoint events, the probability of any event occurring is the sum of the probabilities of each event.
General formula for the union of multiple disjoint events.
The probability of drawing a heart, a spade, or a diamond from a deck of cards.
Useful in complex scenarios involving several mutually exclusive outcomes.
Binomial Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=(kn)pk(1−p)n−k
Calculates the probability of exactly k successes in n independent Bernoulli trials, each with success probability p.
Probability of getting exactly k successes.
Binomial coefficient representing the number of ways to choose k successes from n trials.
Total number of trials.
Number of successful trials.
Probability of success on a single trial.
Probability of failure on a single trial.
If n=5, p=0.6, and k=3, then P(X=3)=(35)(0.6)3(0.4)2≈0.3456.
Used to find the likelihood of a specific number of successes in processes like quality control, clinical trials, and survey analysis.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=∑i=0k(in)pi(1−p)n−i
Calculates the probability of obtaining at most k successes in n trials.
Probability of getting k or fewer successes.
Summation from i=0 to i=k.
All other variables are as previously defined in the PMF.
For n=5, p=0.6, and k=3, P(X≤3) sums the probabilities for k=0 to k=3.
Assessing cumulative probabilities in risk assessment and reliability engineering.
Mean (Expected Value)
Formula:
μ=E[X]=np
Provides the average number of successes expected over n trials.
Mean or expected number of successes.
Expected value of the random variable X.
Total number of trials.
Probability of success on a single trial.
If n=10 and p=0.5, then μ=10×0.5=5.
Predicting average outcomes in repeated experiments like manufacturing defects.
Variance
Formula:
σ2=Var(X)=np(1−p)
Measures the dispersion of the number of successes around the mean.
Variance of the random variable X.
Variance notation.
Total number of trials.
Probability of success on a single trial.
Probability of failure on a single trial.
For n=10 and p=0.5, σ2=10×0.5×0.5=2.5.
Assessing variability in outcomes for quality assurance and forecasting.
Standard Deviation
Formula:
σ=np(1−p)
Represents the average distance of the data from the mean.
Standard deviation of the random variable X.
As previously defined.
With n=10 and p=0.5, σ=10×0.5×0.5≈1.5811.
Used to understand the spread of binomial outcomes in fields like finance and epidemiology.
Poisson Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=k!e−λλk
Calculates the probability of observing exactly k events in a fixed interval when events occur independently at a constant average rate λ.
Probability of exactly k occurrences in the interval.
The probability of zero occurrences.
Mean number of occurrences raised to the k-th power.
Factorial of k, the number of occurrences.
Average rate of occurrence (mean number of events per interval).
Number of occurrences (non-negative integer).
If λ=4 and k=2, then P(X=2)=2!e−4×42≈0.1465.
Modeling the number of rare events over a fixed period, such as the number of incoming calls to a call center in an hour.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=e−λ∑i=0ki!λi
Calculates the probability of observing up to k events in a fixed interval.
Probability of k or fewer occurrences.
Summation from i=0 to i=k.
Average rate of occurrence.
Factorial of i, each term in the sum.
For λ=3 and k=2, P(X≤2)=e−3(0!30+1!31+2!32)≈0.4232.
Assessing cumulative probabilities, such as the probability of observing a certain number of defects in a batch.
Mean (Expected Value)
Formula:
μ=E[X]=λ
Represents the average number of events occurring in the given interval.
Mean or expected value of the distribution.
Expected value operator.
Average rate of occurrence.
If λ=5, then μ=5.
Predicting average outcomes in processes like arrival rates, such as customers entering a store per hour.
Variance
Formula:
σ2=Var(X)=λ
Measures the dispersion of the number of events around the mean.
Variance of the distribution.
Variance operator.
Average rate of occurrence.
If λ=5, then σ2=5.
Assessing variability in event counts for planning and resource allocation.
Standard Deviation
Formula:
σ=λ
Represents the average distance of the data from the mean.
Standard deviation of the distribution.
Average rate of occurrence.
If λ=5, then σ=5≈2.2361.
Understanding the spread of Poisson-distributed outcomes in fields like telecommunications and biology.
Geometric Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=(1−p)k−1p
Calculates the probability that the first success occurs on the k-th trial.
Probability that the first success occurs on trial k.
Probability of success on each trial.
Probability of failure on each trial.
Trial number on which the first success occurs (k=1,2,3,…).
If p=0.3 and k=4, then P(X=4)=(1−0.3)4−1×0.3=(0.7)3×0.3≈0.1029.
Used to model the number of trials needed to get the first success, such as in quality control testing or reliability analysis.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=1−(1−p)k
Calculates the probability that the first success occurs on or before the k-th trial.
Probability that the first success occurs on or before trial k.
Probability of success on each trial.
Probability of failure on each trial.
Trial number (k=1,2,3,…).
For p=0.3 and k=4, P(X≤4)=1−(1−0.3)4=1−(0.7)4≈0.7599.
Assessing the likelihood of achieving the first success within a certain number of trials.
Mean (Expected Value)
Formula:
μ=E[X]=p1
Represents the average number of trials required to achieve the first success.
Mean or expected value of the distribution.
Probability of success on each trial.
If p=0.3, then μ=0.31≈3.3333.
Predicting average trial counts in processes like quality inspections or repeated experiments.
Variance
Formula:
σ2=Var(X)=p21−p
Measures the dispersion of the number of trials around the mean.
Variance of the distribution.
Probability of success on each trial.
Probability of failure on each trial.
If p=0.3, then σ2=(0.3)21−0.3=0.090.7≈7.7778.
Assessing variability in the number of trials needed for the first success.
Standard Deviation
Formula:
σ=p21−p
Represents the average distance of the number of trials from the mean.
Standard deviation of the distribution.
Probability of success on each trial.
Probability of failure on each trial.
If p=0.3, then σ=(0.3)21−0.3≈7.7778≈2.7889.
Understanding the spread of trials needed to achieve the first success.
Negative Binomial Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=(r−1k−1)pr(1−p)k−r
Calculates the probability that the r-th success occurs on the k-th trial.
Probability that the r-th success occurs on trial k.
Binomial coefficient representing the number of ways to arrange r−1 successes in k−1 trials.
Probability of success on each trial.
Probability of failure on each trial.
Trial number on which the r-th success occurs (k=r,r+1,r+2,…).
Number of successes to be achieved.
If p=0.4, r=3, and k=7, then P(X=7)=(26)(0.4)3(0.6)4≈0.0595.
Used to model the number of trials needed to achieve a specified number of successes, such as in reliability testing or sequential analysis.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=Ip(r,k−r+1)
Calculates the probability that the r-th success occurs on or before the k-th trial using the regularized incomplete beta function.
Probability that the r-th success occurs on or before trial k.
Regularized incomplete beta function.
Probability of success on each trial.
Trial number.
Number of successes.
For p=0.4, r=3, and k=7, P(X≤7) can be calculated using the cumulative sum of PMF values or the incomplete beta function.
Assessing the likelihood of achieving r successes within a certain number of trials.
Mean (Expected Value)
Formula:
μ=E[X]=pr
Represents the average number of trials required to achieve r successes.
Mean or expected value of the distribution.
Number of successes.
Probability of success on each trial.
If r=3 and p=0.4, then μ=0.43=7.5.
Predicting average trial counts in processes like quality control or repeated experiments until a certain number of successes are achieved.
Variance
Formula:
σ2=Var(X)=p2r(1−p)
Measures the dispersion of the number of trials around the mean.
Variance of the distribution.
Number of successes.
Probability of success on each trial.
Probability of failure on each trial.
If r=3 and p=0.4, then σ2=(0.4)23×0.6=11.25.
Assessing variability in the number of trials needed for achieving r successes.
Standard Deviation
Formula:
σ=p2r(1−p)
Represents the average distance of the number of trials from the mean.
Standard deviation of the distribution.
Number of successes.
Probability of success on each trial.
Probability of failure on each trial.
If r=3 and p=0.4, then σ=11.25≈3.3541.
Understanding the spread of trials needed to achieve r successes.
Hypergeometric Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=(nN)(kK)(n−kN−K)
Calculates the probability of observing exactly k successes in n draws without replacement from a finite population.
Probability of getting exactly k successes.
Number of ways to choose k successes from K total successes in the population.
Number of ways to choose n−k failures from N−K total failures.
Total number of ways to choose n items from N total items.
Total number of items in the population.
Total number of successes in the population.
Number of draws or trials.
Number of observed successes.
If N=50, K=10, n=5, and k=2, then P(X=2)=(550)(210)(340).
Used in scenarios involving sampling without replacement, such as quality control inspection or lotteries.
Mean (Expected Value)
Formula:
μ=E[X]=n(NK)
Represents the average number of successes expected in n draws.
Mean or expected value of the distribution.
Number of draws.
Total number of successes in the population.
Total number of items in the population.
If N=50, K=10, and n=5, then μ=5(5010)=1.
Predicting the average number of successes in random sampling without replacement.
Variance
Formula:
σ2=Var(X)=n(NK)(NN−K)(N−1N−n)
Measures the dispersion of the number of successes around the mean.
Variance of the distribution.
Number of draws.
Total number of successes in the population.
Total number of items in the population.
Total number of failures in the population.
Number of items not drawn.
Adjustment for finite population.
If N=50, K=10, and n=5, then σ2=5(5010)(5040)(4945)≈0.9184.
Assessing variability in outcomes when sampling without replacement.
Standard Deviation
Formula:
σ=n(NK)(NN−K)(N−1N−n)
Represents the average distance of the number of successes from the mean.
Standard deviation of the distribution.
All other variables are as previously defined.
Using the previous variance example, σ=0.9184≈0.9583.
Understanding the spread of the number of successes in sampling without replacement.
Probability Of At Least $k$ Successes
Formula:
P(X≥k)=1−∑i=0k−1P(X=i)
Calculates the probability of observing at least k successes in n draws.
Probability of k or more successes.
Sum of probabilities of observing fewer than k successes.
PMF evaluated at i successes.
To find P(X≥2), compute 1−[P(X=0)+P(X=1)].
Assessing the likelihood of meeting or exceeding a certain number of successes.
Calculates the probability of obtaining a specific combination of counts (x1,x2,…,xk) in n independent trials, where each trial can result in one of k possible outcomes.
Probability of observing counts x1,x2,…,xk for outcomes 1 to k respectively.
Total number of independent trials.
Number of times outcome i occurs, where ∑i=1kxi=n.
Probability of outcome i on a single trial, where ∑i=1kpi=1.
Factorial of n.
Factorial of xi.
If n=5, k=3, p1=0.2, p2=0.5, p3=0.3, and we observe counts x1=1, x2=2, x3=2, then P=1!2!2!5!(0.2)1(0.5)2(0.3)2≈0.135.
Used in experiments with more than two possible outcomes per trial, such as dice rolls, categorical data analysis, or genetic allele frequencies.
Mean (Expected Value) Of Each Outcome
Formula:
E[Xi]=npi
Represents the average number of times outcome i is expected to occur in n trials.
Expected value of the count for outcome i.
Total number of trials.
Probability of outcome i.
For n=10 and p2=0.4, E[X2]=10×0.4=4.
Predicting average counts of each category in repeated experiments.
Variance Of Each Outcome
Formula:
Var(Xi)=npi(1−pi)
Measures the dispersion of the count for outcome i around its mean.
Variance of the count for outcome i.
Total number of trials.
Probability of outcome i.
Probability of not observing outcome i.
For n=10 and p2=0.4, Var(X2)=10×0.4×0.6=2.4.
Assessing variability in counts for each category.
Covariance Between Outcomes
Formula:
Cov(Xi,Xj)=−npipj
Measures the degree to which counts of outcomes i and j vary together.
Covariance between counts of outcomes i and j.
Total number of trials.
Probabilities of outcomes i and j, respectively.
For n=10, p1=0.2, p2=0.5, Cov(X1,X2)=−10×0.2×0.5=−1.
Understanding the negative dependence between category counts in multinomial experiments.
Quantifies the linear relationship between counts of outcomes i and j.
Correlation coefficient between counts of outcomes i and j.
Covariance between counts of outcomes i and j.
Variances of counts of outcomes i and j, respectively.
Probabilities of outcomes i and j.
Using previous values, ρ12=0.2×0.8×0.5×0.5−0.2×0.5≈−0.577.
Assessing the strength of the relationship between category counts.
Discrete Uniform Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=b−a+11
Calculates the probability of the discrete random variable X taking on a specific integer value k within the range from a to b, where all outcomes are equally likely.
Probability that X equals k.
Minimum integer value in the distribution's range.
Maximum integer value in the distribution's range.
An integer within the range a≤k≤b.
Total number of possible integer values in the range.
If a=1 and b=6 (like a fair die), then P(X=3)=61.
Modeling situations where a finite number of outcomes are equally likely, such as rolling a fair die or selecting a random card from a deck.
Mean (Expected Value)
Formula:
μ=E[X]=2a+b
Represents the average value of the distribution, calculated as the midpoint of the range.
Mean or expected value of the distribution.
Minimum integer value.
Maximum integer value.
For a=1 and b=6, μ=21+6=3.5.
Determining the average expected outcome in uniformly random scenarios.
Variance
Formula:
σ2=Var(X)=12(b−a+1)2−1
Measures the dispersion of the values around the mean.
Variance of the distribution.
Minimum integer value.
Maximum integer value.
Total number of possible integer values.
For a=1 and b=6, σ2=12(6−1+1)2−1=1236−1=1235≈2.9167.
Assessing the spread of outcomes in uniformly random scenarios.
Standard Deviation
Formula:
σ=12(b−a+1)2−1
Represents the average distance of the values from the mean.
Standard deviation of the distribution.
All other variables are as previously defined.
Using the previous variance example, σ=2.9167≈1.7078.
Understanding the spread of uniformly distributed discrete outcomes.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=b−a+1k−a+1 for k=a,a+1,…,b
Calculates the probability that the random variable X is less than or equal to a specific value k.
Cumulative probability up to and including k.
Minimum integer value.
Maximum integer value.
An integer within the range a≤k≤b.
For a=1, b=6, and k=3, P(X≤3)=6−1+13−1+1=63=0.5.
Determining cumulative probabilities in uniformly random discrete events.
Negative Hypergeometric Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=(KN)(r−1k−1)(K−rN−k)
Calculates the probability that the r-th success occurs on the k-th trial when sampling without replacement from a finite population.
Probability that the r-th success occurs on trial k.
Number of ways to choose r−1 successes in the first k−1 trials.
Number of ways to choose the remaining K−r successes from the remaining N−k items.
Total number of ways to choose K successes from N items.
Total number of items in the population.
Total number of successes in the population.
Trial number on which the r-th success occurs (k=r,r+1,…,N−K+r).
Number of successes to be achieved.
If N=20, K=5, r=2, and k=4, then P(X=4)=(520)(13)(316)≈0.2381.
Used to model the number of trials needed to achieve r successes without replacement, such as in card games or quality inspections.
Mean (Expected Value)
Formula:
μ=E[X]=K+1r(N+1)
Represents the average number of trials required to achieve r successes when sampling without replacement.
Mean or expected value of the distribution.
Number of successes to be achieved.
Total number of items in the population.
Total number of successes in the population.
If N=20, K=5, and r=2, then μ=5+12×21=7.
Predicting average trial counts in processes like drawing specific cards from a deck without replacement.
Variance
Formula:
σ2=(K+1)2(K+2)r(N+1)(N−K)(N−r)
Measures the dispersion of the number of trials around the mean.
Variance of the distribution.
Number of successes to be achieved.
Total number of items in the population.
Total number of successes in the population.
Number of failures in the population.
Number of items not yet drawn after achieving r successes.
If N=20, K=5, and r=2, then σ2=(5+1)2×72×21×15×18≈5.7143.
Assessing variability in the number of trials needed for achieving r successes without replacement.
Standard Deviation
Formula:
σ=(K+1)2(K+2)r(N+1)(N−K)(N−r)
Represents the average distance of the number of trials from the mean.
Standard deviation of the distribution.
All other variables are as previously defined.
Using the previous variance example, σ=5.7143≈2.3917.
Understanding the spread of trials needed to achieve r successes without replacement.
Cumulative Distribution Function (CDF)
Formula:
P(X≤k)=1−(kN)(k−rN−r)(r−1r−1)
Calculates the probability that the r-th success occurs on or before the k-th trial.
Probability that the r-th success occurs on or before trial k.
Number of ways to choose k−r failures from N−r failures.
Total number of ways to choose k items from N items.
Trial number.
For N=20, K=5, r=2, and k=7, compute P(X≤7) accordingly.
Assessing the likelihood of achieving r successes within a certain number of trials when sampling without replacement.
Logarithmic Distribution
Probability Mass Function (PMF)
Formula:
P(X=k)=−ln(1−p)1kpk
Calculates the probability of the discrete random variable X taking on the value k, where k represents the number of failures required to get the first success in a logarithmic series.
Probability that X equals k.
Parameter of the distribution, such that 0<p<1.
A positive integer (k=1,2,3,…).
Natural logarithm of 1−p, normalizing constant.
If p=0.5 and k=3, then P(X=3)=−ln(1−0.5)13(0.5)3≈0.0625.
Modeling species abundance in ecology, or the number of times a rare event occurs.
Mean (Expected Value)
Formula:
μ=E[X]=(1−p)ln(1−p)−p
Represents the average value of the distribution.
Mean or expected value of the distribution.
Parameter of the distribution.
Natural logarithm of 1−p.
Complement of parameter p.
If p=0.5, then μ=(0.5)(−0.6931)−0.5≈1.4427.
Determining the average number of occurrences in phenomena following a logarithmic distribution.
Variance
Formula:
σ2=Var(X)=(1−p)2[ln(1−p)]2−p(p+ln(1−p))
Measures the dispersion of the values around the mean.
Variance of the distribution.
Parameter of the distribution.
Natural logarithm of 1−p.
Complement of parameter p.
If p=0.5, calculate σ2 accordingly.
Assessing the spread of outcomes in logarithmically distributed data.
Standard Deviation
Formula:
σ=Var(X)
Represents the average distance of the values from the mean.
Standard deviation of the distribution.
Variance of the distribution.
Calculate σ using the variance from the previous example.
Understanding the spread of logarithmically distributed discrete outcomes.
Generating Function
Formula:
GX(s)=ln(1−p)ln(1−ps)
Generates the probabilities of the distribution and can be used to find moments.
Probability generating function of the distribution.
Parameter of the distribution.
Dummy variable used in generating functions.
Natural logarithm of 1−ps.
Natural logarithm of 1−p.
For p=0.5, compute GX(s) accordingly.
Used in advanced probability to derive properties of the distribution.