Prof. Dr. Markus Meier
Leibniz Institute for Baltic Sea Research Warnemünde (IOW)
E-Mail: markus.meier@io-warnemuende.de
Probability density and distribution#
Probability density function and important parameters
Different probability distributions
Probability density function and important parameters#
Probability density function#
let
be a continous (not discrete!) variable that takes values in , for example temperature, probability density function of an event X (i.e. T=10°C) is defined as a continous function on with the following three attributes:
Question: What is the unit of the pdf?
Answer: [ ] = [x] .Question: What is the integral of the pdf? Answer: The cumulative distribution function.
Cumulative distribution function#
Cumulative distribution function for an event X is a montonously increasing, non-dimensional function F_X(x) on
defined as:
which is equivalent to:
consequently the probability of the event
to be inside the range of is:
Expectation #
the expectation of a given pdf weighs is with
in the integral:
two attributes of the expectation are:
Central moments #
k-th moment of a continous random variable X:
k-th central moment of a continous random variable X:
example: anomalies with
mean seasonal cyclemean
: location parametervariance:
standard deviation:
Chebyshev’s inequality:
Skewness #
is a measure of the asymmetry of a distribution: symmetric for
, scaled version of the third central moment, non-dimensional shape parameter
Kurtosis #
is a measure of the peakedness of a distribution: a normal distribution (will be explained later this lecture) has
, scaled and shifted version of the fourth central moment, non-dimensional shape parameter
Examples#
summer sea level at Kieler Förde,
, , ,
probability densities of some measured variables
P-quantiles#
mean and variance are affected by the tail ends of the pdf (likelihood of extreme values), but p-quantiles
are insensitive to extreme values.p quantile of 0.3 means that 30% of the x values are below this threshold
median m
is the 50%-quantile: half of the distribution lays above and the other half below m .
let’s look at the p-quantiles of the log-normal distribution in Figure 5 to get an idea. note the difference of mean and median!
Different probability distributions#
Uniform distribution#
symmetric and less peaked than the normal distribution:
with the cumulative distribution function:
exercise: calculate
of the uniform distribution
solutions: , , , ,
Normal (Gaussian) distribution#
most physical quantities are nearly normal distributed
no skewness or kurtosis:
no analytical form of cdf, approximation:
central limit theorem states: If
is an infinite series of independent and identically distributed random variables with and then the average is asymptotically normal distributed. That is:
a larger sample size reduces the standard deviation as of:
Log-normal distribution#
distribution of positive definite quantities such as rainfall, wind speed
with the median value
and
exercise: derive a general for the k-th central moment of the distribution
solution:
-distribution#
sum of k independent squared
random variables, k number of degrees of freedom, application for the pdfs of variance estimates:
with
it has handy attributes:
Student’s t-distribution#
application for testing the significance of the differences in the means. be
a test variable with , if A and B are independent random variables such that
the t-distribution can be written as:
using the
-function (28) the distribution can also be written as:
t-test?
Fisher-F-distribution#
application for testing the significance of the differences in the variance. for
-distributed and :
the F-distribution is given by:
alternatively the probsbility density of the F-distribution is also given by:
Summary of theoretical distributions#
Continous random vectors, multi-variate data#
example: vectors X temperature and Y sea level pressure: