Unit 1: Introduction to Statistics

Statistics is the science of collecting, organising, summarising and analysing data to draw conclusions from it. It can be divided in two branches:

As with many disciplines, the scientific method can be easily applied to statistics.

Basic Concepts

A population is a set of elements in which we’re interested. A sample is a subset of a population.

A variable is a feature of a population. Variables can measure both quantities and qualities. The set of observations for a specific variable in a given population is called data.

Measures

Measures of Central Tendency

A measure of central tendency is a value in a set of data that can be described as being in the centre of the (sorted) data. The most common measures of central tendency are the arithmetic mean, the weighted mean, the median and the mode.

In the case of the arithmetic mean, it is also possible to calculate it when we have a table of frequencies or groups of data with class marks. Similarly, we can calculate the median for a table of frequencies.

Measures of Position

A quantile is a value that serves as a measure of position. They can be seen as the value up to which a random variable accumulates a given density. Percentiles, quartiles and deciles are specific cases of quantiles.

Measures of Dispersion

A measure of dispersion is a value that helps us describe how disperse the mass of a variable is. Variance, standard deviation, mean deviation and range are all examples of measures of dispersion.

Measures of Shape

A measure of shape is a value that helps us describe the shape that a variable’s mass has. Moments, skewness, kurtosis and excess kurtosis are examples of measures of shape.

Measures of Association

A measure of association is a value that quantifies the relationship between two or more variables. Covariance and Pearson’s correlation coefficient are examples of measures of association.

Graphical Representations

Statistical data can be represented graphically in many types of graphs, which include, but are not limited to:

  • Bar charts, where classes of a variable are associated with a frequency bar.
  • Histograms, where values are ordered.
  • Dispersion graphs, which show the relation between two variables.
  • Bubble graphs, which show the relation between three variales.
  • Time series, which show the evolution of a variable across time.
  • Box plots, which summarise a series of numerical data with the help of quartiles.

Unit 2: Methods for Obtaining Functions of Random Variables

Probability Theory

(Not so Brief) Probability Theory Review

A brief review on probability theory was seen, which included the concepts of random variable (discrete and continuous), density function, distribution function, random vector, and moment-generating function, to name just a few.

I created no new notes for these concepts, since I had already seen all of them during my Probabilidad course (🏫). I did improve and edit some of them, however.

When defining a random variable in terms of another, we can find its density and distribution function by using one of two methods.

Collections of iid Random Variables

We say that a collection of random variables is iid when all of them are mutually independent and have the same probability distribution.

More specifically, when sampling a population and describing the random variables whose realisations we have obtained as samples, we refer to a collection of iid random variables as a random sample.

When ordering the random variables in a random sample according to their realisations, we can define the concept of order statistics.

More Probability Distributions

When having a collection of random variables that follow a standard normal distribution, the random variable that results from adding their squares up follows what we call a chi-squared distribution (denoted, of course, with the Greek letter chi ).

Student’s t-distribution describes the distribution of a random variable that results from operating a normally distributed random variable with a chi-squared distributed one.

Similarly, when operating two random variables that follow a chi-squared distribution, the resulting random variable has an F distribution.

Unit 3: Sampling Distributions

Random Samples, Sample Statistics and Parameters

When sampling a population and describing the random variables whose realisations we have obtained as samples, we refer to a collection of iid random variables as a random sample.

A sample statistic (”estadístico” in Spanish) is a random variable that results from passing a random sample through a function. The corresponding distributions of such random variables are called sampling distributions.

A parameter is a numerical characterisation of a population that describes partially or completely its probability distribution. The set of all possible values that a parameter can have is called its parameter space.

Expected Value, Variance and Moment-Generating Function Theorems

When defining a random variable as the sum of the random variables in a random sample multiplied by some scalar each, its moment-generating function can be easily determined.

When we know the expected value and variance of the random variables of our random sample, then obtaining the expected value and variance of the mean of our random sample is easy.

Central Limit Theorem

The central limit theorem is an important theorem that allows us to use a standard normal distribution to approximate a sum of iid random variables with any distribution.

Sample Variance Theorems

Given a normally distributed random sample, we can use a chi-squared distribution to work with its variance. We can do something similar even when we don’t know the expected value of the random variables, albeit with a slightly different sample variance.

Unit 4: Point and Interval Estimation

Estimators

Given a random sample characterised by a parameter, an estimator is a sample statistic that aims to estimate (i.e. give an approximation of) such a parameter.

An estimator can be:

  • Unbiased when its expected value is equal to the parameter it estimates.
  • Asymptotically unbiased (in the case of a sequence of estimators) when, as the size of the sequence increases, the expected values converge to the estimated parameter.
  • More efficient than others when their variance is smaller.
  • Consistent (in the case of a sequence of estimators) when, as the size of the sequence increases, they converge to the estimated parameter.

The bias of an estimator is the difference between its expected value and the parameter it estimates. The mean squared error (MSE) is the expected value of the square of this difference.

Obtaining Estimators

We can use several methods to obtain estimators for a given parameter:

  • The moments method consists of equating the population moments with their corresponding sample moments, and then solving the resulting equation(s).
  • The maximum likelihood method consists of maximising the random sample’s likelihood function.
    • As its name suggests, this method uses a random sample’s likelihood function, which is the product of the density functions of each random variable in the sample.

Cramér-Rao Lower Bound

Given an estimator , the Cramér-Rao lower bound provides a lower bound for .

An estimator is said to be a uniformly minimum-variance unbiased (UMVUE) estimator when (1) it’s unbiased and (2) its Cramér-Rae lower bound is equal to .