Week 5: Estimation & Testing
Model Estimation and Testing
Week Learning Objectives
By the end of this module, you will be able to
- Describe, conceptually, what likelihood function and maximum likelihood estimation are
- Describe the differences between maximum likelihood and restricted maximum likelihood
- Conduct statistical tests for fixed effects, and use the small-sample correction when needed
- Use the likelihood ratio test to test random slopes
- Estimate multilevel models with the Bayesian/Markov Chain Monte Carlo estimator in the
brms
package
Task List
- Review the resources (lecture videos and slides)
- Complete the assigned readings
- Snijders & Bosker ch 4.7, 6, 12.1, 12.2
- Attend the Tuesday session and participate in the class exercise
- (Optional) Fill out the early/mid-semester feedback survey on Blackboard
Lecture
Slides
Overview
Estimation methods
Likelihood function
Estimation methods for MLM
Likelihood ratio test (LRT) for fixed effects
The LRT has been used widely across many statistical methods, so itβs helpful to get familiar with doing it by hand (as it may not be available in all software in all procedures).
\(F\) test with small-sample correction
For more information on REML and K-R, check out
- McNeish, D. (2017). Small sample methods for multilevel modeling: A colloquial elucidation of REML and the Kenward-Roger correction.
LRT for random slope variance
Using brms
Check your learning: Using R, verify that, if \(\mu = 10\) and \(\sigma = 8\) for a normally distributed population, the probability (joint density) of getting students with scores of 23, 16, 5, 14, 7.
Check your learning: Using the principle of maximum likelihood, the best estimate for a parameter is one that
More about maximum likelihood estimation
If \(\sigma\) is not known, the maximum likelihood estimate is \[\hat \sigma = \sqrt{\frac{\sum_{i = 1}^N (Y_i - \bar X)^2}{N}},\] which uses \(N\) in the denominator instead of \(N - 1\). Because of this, in small samples, maximum likelihood estimates tend to be biased, meaning that, on average, it tends to underestimate the population variance.
One useful property of maximum likelihood estimation is that the standard error can be approximated by the inverse of the curvature of the likelihood function at the peak. The two graphs below show that with a larger sample, the likelihood function has a higher curvature (i.e., steeper around the peak), which results in a smaller estimated standard error.