Learning Outcomes Week 1
- Understand that machine learning combines data with assumptions to make predictions
- Understand that probability provides a calculus of uncertainty for us to deal with unknowns.
- Understand the definition of entropy
- Understand that the KL-divergence is an asymmetric measure of similarity between probability distributions
- Understand sample based approximations
- Be able to prove that maximum likelihood solution is approximately minimizing the KL-divergence
- Be able to derive an error function from the likelihood of a single data point.
- Independence of data points (data is i.i.d.)
- Logarithm is monotonic
- In optimization, by convention, we minimize so take negative.
Learning Outcomes Week 2
- Consolidate understanding of stages of a basic probabilistic machine learning:
- Write down model.
- Make an assumption about the errors.
- Use combination of mathematical model and error assumptions to write down a likelihood
- Maximize the likelihood with respect to the parameters of the model
- Use the resulting model to make predictions.
- Understand the principles of using gradient methods to find a fixed point equation to maximize a likelihoood.
- Understand the weakness of coordinate descent methods when parameters are correlated.
- Understand the advantages of using multivariate calculus to maximize the likelihood in linear regression.
- Understand how basis functions can be used to go from linear models to non-linear models.
Learning Outcomes Week 3
- Understand the challenge of model selection.
- Understand the difference between training set, test set and validation set.
- Understand and be able to apply appropriately the following approaches to model validation:
- hold out set,
- leave one out cross validation,
- k-fold cross validation.
- Be able to identify the type of error that arises from bias and the type of error that arises from variance.
- Be able to distinguish between different types of uncertainty: aleatoric and epistemic. Be able to give examples of each type.
- Be able to derive Bayes rule' from the product rule of probability.
- Understand the meaning of the terms prior, posterior and marginal likelihood
- Be able to identify these terms in Bayes' rule.
- Be able to describe what each of these terms represents (belief before observation, belief after observation, relationship between belief and observation, the model score.)
- Understand how to derive the marginal likelihood from the likelihood and the prior.
- Understand the difference between the frequentist approach and the Bayesian approach, i.e. that in the Bayesian approach parameters are treated as random variables
- Be able to derive the maths to perform a simple Bayesian update on the offset parameter of a regression problem.
Learning Outcomes Week 5
- Understand the principal of integrating parameters and how to use Bayes rule to do so.
- Understand the role of the prior distribution.
- In multivariate and univariate Gaussian examples, be able to combine the prior with the likelihood to form a posterior distribution..
- Recognise the role of the marginal likelihood and know its form for Bayesian regression under Gaussian priors.
- Be able to compute the expected output of the model and its covariance using the posterior distribution and the formula for the function.
- Understand the effect of model averaging and its advantages when making predictions including:
- Error bars
- Regularized prediction (reduces variance)
Learning Outcomes Week 7
- Understanding how the marginal likelihood in a Gaussian Bayesian regression model can be computed using properties of the multivariate Gaussian.
- Understanding that Bayesian Regression Models put a joint Gaussian prior across the data.
- Understanding that we can specify the covariance function of that prior directly.
- Understanding that Gaussian process models generalize basis function models to allow infinite basis functions.
This document last modified Friday, 24-Jan-2014 08:12:02 UTC