A list of likely exam questions.
- Describe a polynomial fit.
- What is a cost function (or loss function, or error function)?
- What do we mean by "overfitting"?
- What is a training data set? and what is a test data set?
- Describe "linear prediction".
- Discuss the difference between the definition of probability
in probability space and empirical probability.
- "Prove" Bayes' theorem and discuss its relevance in inference
problems.
- Prove Chebyshev's inequality and use it to find the weak law
of large numbers.
- Describe the link between probability generating functions and
momenta of discrete distributions.
- Find the probability generating function of a sum of random
variables.
- Find the probability generating function of common
distributions, Poisson, Bernoulli, Binomial.
- Find the characteristic function of the Gaussian distribution.
- Give an outline of the proof of the Central Limit Theorem.
- Describe the geometric distribution.
- Describe the negative binomial distribution.
- Describe the hypergeometric distribution.
- Describe the elementary model of cancer onset vs. age.
- Demonstrate the formula for the detected rate vs. true rate
for paralizable and non-paralizable counters.
- Prove the formula for the coincidence rate of multiple
detectors.
- Find the high-rate limit of the Poisson distribution.
- Describe in detail the multivariate Gaussian distribution.
- Describe the gamma distribution and prove the formulas for
expectation value and variance.
- Prove that the correlation coefficient is lies in the (-1,+1)
interval.
- Find the probability distribution of the sample mean assuming
an exponential distribution for individual measurements.
- What is a box plot?
- What is a kernel density estimate?
- What is an empirical probability density?
- Discuss the algorithmic generation of pseudo-random numbers.
- Describe the transformation method to generate arbitrarily
distributed pseudo-random numbers
- Describe the acceptance-rejection method and provide an
elementary proof of its validity.
- Discuss the statistical uncertainty of Monte Carlo results.
- Describe the statistical bootstrap.
- Describe the principle of maximum likelihood (MaxL).
- Discuss the application of MaxL to normally distributed data,
to estimate mean and/or standard deviation.
- Discuss the application of MaxL to data extracted from a
Poisson distribution to estimate the mean.
- Prove the Bartlett identities.
- Prove the Cramer-Rao bound
- Introduce Fisher information with one parameter and with
multiple parameters.
- Define the Shannon information.
- Prove the Jensen inequality.
- Use the Jensen inequality to prove that the KL divergence is
non-negative.
- Give examples of application of the KL divergence to simple
distributions.
- Describe credibility intervals in a Bayesian context.
- Outline the Neyman construction for confidence intervals.
- Describe the method based on likelihood to estimate the
confidence intervals of model parameters.
- How to use a statistic to discriminate between null and
alternative hypothesis.
- Discuss the relevance of the chi-square distribution in
hypothesis testing
- Prove the Neyman-Pearson lemma.
- What are p-values and how do you use them in multiple tests of
the null hypothesis?
- Describe the principal component method.
December 2025.