Skip to content

edrubin/EC524W26

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

180 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

EC 524/424, Winter 2026

Welcome to Economics 524 (424): Prediction and machine-learning in econometrics, taught by Ed Rubin and Jose Rojas Fallas.

Schedule

Lecture Tuesdays and Thursdays, 10:00a-11:20a (Pacific), 105 Esslinger

Lab Friday, 10:00a–10:50a (Pacific), 072 PLC

Office hours

  • Ed Rubin Tu. 3:30p–4:30p (PLC 530)
  • Jose Rojas Fallas We. 2p–3p (PLC 525)

Syllabus

Syllabus

Books

Required books

Suggested books

Lecture notes

Note: Links to topics that we have not yet covered lead to older slides. I will update links to the new slides as we work our way through the term/slides.

000 - Overview (Why predict?)

  1. Why do we have a class on prediction?
  2. How is prediction (and how are its tools) different from causal inference?
  3. Motivating examples

Formats .html | .pdf | .rmd

Readings Introduction in ISL

001 - Statistical learning foundations

  1. Why do we have a class on prediction?
  2. How is prediction (and how are its tools) different from causal inference?
  3. Motivating examples

Formats .html | .pdf | .rmd

Readings

Supplements Unsupervised character recognization

002 - Model accuracy

  1. Model accuracy
  2. Loss for regression and classification
  3. The variance-bias tradeoff
  4. The Bayes classifier
  5. KNN

Formats .html | .pdf | .rmd

Readings

  • ISL Ch2–Ch3
  • Optional: 100ML Preface and Ch1–Ch4

003 - Resampling methods

  1. Review
  2. The validation-set approach
  3. Leave-out-out cross validation
  4. k-fold cross validation
  5. The bootstrap

Formats .html | .pdf | .rmd

Readings

  • ISL Ch5
  • Optional: 100ML Ch5

004 - Linear regression strikes back

  1. Returning to linear regression
  2. Model performance and overfit
  3. Model selection—best subset and stepwise
  4. Selection criteria

Formats .html | .pdf | .Rmd

Readings

  • ISL Ch3
  • ISL Ch6.1

In between: tidymodels-ing

005 - Shrinkage methods

(AKA: Penalized or regularized regression)

  1. Ridge regression
  2. Lasso
  3. Elasticnet

Formats .html | .pdf | .Rmd

Readings

  • ISL Ch4
  • ISL Ch6

006 - Classification intro

  1. Introduction to classification
  2. Why not regression?
  3. But also: Logistic regression
  4. And maximum likelihood estimation
  5. Assessment: Confusion matrix, assessment criteria, ROC, and AUC

Formats .html | .pdf | .Rmd

Readings

  • ISL Ch4
  • ISL Ch6

Bonus Two nice interactive visualizations of gradient descent (and related algorithms)—and a mildly related game.

007 - Decision trees

  1. Introduction to trees
  2. Regression trees
  3. Classification trees—including the Gini index, entropy, and error rate

Formats .html | .pdf | .rmd

Readings

  • ISL Ch8.1–Ch8.2

008 - Ensemble methods

  1. Introduction
  2. Bagging
  3. Random forests
  4. Boosting

Formats .html | .pdf | .rmd

Readings

  • ISL Ch8.2

009 - Support vector machines

  1. Hyperplanes and classification
  2. The maximal margin hyperplane/classifier
  3. The support vector classifier
  4. Support vector machines

Formats .html | .pdf | .rmd

Readings

  • ISL Ch9

010 - Unsupervised learning, dimensionality reduction, and image classification

  1. MNIST dataset (machines with vision)
  2. K-means clustering
  3. Principal component analysis (PCA)
  4. UMAP

Readings

  • ISL Ch12

Formats .html | .pdf | .rmd

Also: An older notebook... .html | .qmd

011 - A quick introduction to neural networks

  1. anatomy of a single neuron
  2. logistic regression as a one-neuron network
  3. activation functions and hidden layers
  4. a brief overview of training
  5. the return of MNIST

Readings

Formats .html | .pdf | .rmd

Interesting ML/AI applications

Starting in week 2, you will submit brief write-ups for machine-learning/AI applications that you find interesting. These nine write-ups together will count for a single "project" grade.

Submit by Monday at 11:59p (Pacific) each week (on Canvas):

  • What is the application?
  • What kind of prediction problem is it?
  • What ML/AI methods are they using?
  • Why is this interesting/useful?
  • URL/file for the article or post.

The "coolest" applications will be highlighted in class and will receive extra credit.

Week 2

Week 3

Week 4

Week 5 Insurance week!

Week 6 ML-driven matching.

Week 7 Audits.

Week 8 More (nuanced?) audits and outlier detection.

Projects

Past, present, and future projects.

Example Using tidymodels (and tidyverse) with the Chicago housing data.

000 An introduction to prediction and resampling

001 Cross validation and penalized regression

002 Classification

003 Trees, ensembles, and boosting

004 Prediction finale (cancelled)

Class projects

Class project 01: Application
Selected topic due by 30 January 2026
Project due 04 March 2026

Class project 02: Extension
Selected topic due by 13 February 2026
Project due 11 March 2026

Final exam

In-class exam: Tuesday (17 March 2026) at 8:00a–10:00a
Note: Some previous years had a take-home portion of the final exam. This year, we will only have an in-class exam.

Prep materials
Previous in-class exams: 2023 | 2024 | 2025
Previous take-home exam: 2023 | 2024
Note: We will not provide keys.

Lab notes

Approximate/planned topics... or at least for reference...

00 - RStudio Review

  1. General "best practices" for coding
  2. Working with RStudio
  3. The pipe (%>%)
  4. Cleaning and Kaggle follow up

Formats .html | .pdf | .Rmd

01 - Workflow and Resampling

  1. Install Quarto. Follow this link, download the installer for your operating system, and follow the instructions to install Quarto
  2. Download (and unzip) the Lab Files
  3. Create a project in Rstudio in a separate folder
  4. Copy/move the Lab Files to a folder dedicated to this lab
  5. Open the Quarto document in Rstudio and follow the instructions

Formats .html | .pdf | .qmd

02 - Introduction to tidymodels

  1. Download the Lab File
  2. We will learn about cleaning data quickly and efficiently with tidymodels

Formats .html

03 - More tidymodels

04 - Classification

Additional resources

Jobs

I wrote a very short guide to finding a job.

For programming-related jobs, get some practice on

R

SQL

Data Science

Spatial data

About

Masters-level applied econometrics course—focusing on prediction—at the University of Oregon (EC424/524 during Winter quarter, 2026). Taught by Ed Rubin and Jose Rojas Fallas.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors