Skip to content

Latest commit

 

History

History
177 lines (122 loc) · 9.45 KB

TWiML_schedule.md

File metadata and controls

177 lines (122 loc) · 9.45 KB

Syllabus and Schedule

This schedule is subject to adjustment. Please check frequently to stay updated.

Saturday webinar meeting time

The weekly webinar will occur at 11AM EST (4PM GMT). Readings should be read prior to Saturday webinar. This is a discussion hour focused on what was covered in the lecture tha week.

Homework schedule

Students can turn in homework assignes any time for grading. However, it is recommended you work on the homeworks and submit them along the following schedule.

Homework Date Assigned Date Due
HW1 February 10 March 01
HW2 March 02 March 22
HW3 March 23 April 05
HW4 April 06 April 26

Submit your homeworks online.

Section 1: Refactored-thinking for Machine Learning and causality

Causality and Model-based Machine Learning:

February 8, 15, and 22, 2020

Many applied data scientists and machine learning engineers have a bias towards curve-fitting, overthinking training, and under-thinking how the data is generated.
After this section, you will have unlearned these biases, and have acquired new mental models for applied data science and machine learning. While we perform this mental refactoring, we will graft on a high-level understanding of causality in the context of machine learning. This will lay the foundation for the rest of the course. But more importantly, you'll have a mental model that will increase your ROI on your future self-study. For this reason, if you were to drop the course after this section, you'd still be ahead of your peers.

Topics

  • Thinking about and modeling the data generating process
  • Model iteration through falsification
  • Directed graphs, causalility, and anti-causal machine learning
  • Deep causal generative models with plate models and tensor programming
  • Causal probabilistic programming
  • Examples from natural and social science

Tutorials

Do Causality like a Bayesian

February 29, 2020

You probably already know about and have applied Bayes rule, or you have at least heard of it. In this section, you will go beyond Baye's rule to acquiring a Bayesian mental model for tackling machine learning problems, and building learning agents that drive decision-making in organizations.

Topics

  • Primer on Bayesian machine learning
  • Communication theory and Bayes
  • Bayesian notation
  • Independence of cause and mechanism
  • Bayesian supervised learning case study
  • Bayesian decision-making
  • Modeling uncertainty in Bayesian models

The Refactored Thinking in Machine Learning Course ends here

Section 2: Core elements of causal inference

How to speak graph, or DAG that's a nice model!

Dates: March 7 and 14, 2020

Graphs provide a language for composing, communicating, and reasoning about generative models, probability, and causality. In this section, you will learn this language. You will have the ability to use graph algorithms to describe and reason about the data's probability distributions.

Topics

  • DAGs, joint probability distributions, and conditional independence
  • D-separation, V-structures/colliders, Markov blanket
  • Markov property and disentangling joint probability
  • Faithfulness, causal sufficiency, and causal minimality
  • Model iteration through falsification
  • Ancestral graphs
  • Markov equivalence
  • Model iteration through falsification

Tutorials

The Tao of Do; Modeling and Simulating Causal Interventions

Dates: March 21 and 28, 2020

An intervention is an action by humans or learning agents that change the data generating process, and thus the distribution underlying the training data. If a machine learning model can predict the outcome of an intervention, it is by definition a causal model. Even the most cutting-edge deep learning models can predict the outcomes of interventions unless they are also causal models.

After this section, students will be able to build their first causal generative machine learning model using a deep learning framework.

Topics

  • Observation vs intervention, and the intervention definition of causality
  • Types of interventions
  • Using interventions to falsify and improve models
  • "do"-notation
  • Intervention prediction in simulation models
  • Interventions as graph mutilation and program transforms
  • Breaking equivalence with interventions
  • Simulating causal effects and potential outcomes
  • Implementation examples from forecasting

Tutorial

Section 3: Applied Causal Inference; Identication and Estimation of Causal Effects from Data

The modern practice of causal inference, particularly in the tech industry, is about estimating causal effects -- i.e. quantification of how much a cause affects an outcome. After this section, you will be able to explain to colleagues when estimation is impossible even when they think they can crack it with enough data or a clever algorithm. You will be able to stand your ground in discussions about causality with Ph.D. statisticians and economists at top tech companies. You will have mastered the programmatic causal effect estimation. You will have gained the foundation needed to go deep into standard estimation methods used in practice.

Dates: April 4, 11 and 18, 2020

Topics

  • Why we care about estimating causal effects
  • Defining "confounding" with DAGs
  • Simpson's Paradox, Monte Hall problem, Berkson's Paradox
  • Statistics of causal effects: the estimand, the estimator, and the estimate
  • Identification: Why causal effect inference is hard no matter how much data you have
  • What is the "do"-calculus?
  • Potential outcomes and individual treatment effects
  • Valid adjustment sets for causal effect estimation
  • Single world intervention graphs
  • Ignorability and SUTVA
  • Introduction to the DoWhy library
  • Statistical estimation methods: G-formula, propensity matching, instrumental variables, inverse probability weighting, and more.

Tutorials

Section 4: Counterfactual machine learning

Dates: April 25 and May 2, 2020

Counterfactual reasoning sounds like "I chose company A, and now I'm miserable but had I worked for company B, I would have been happy." We make decisions and observe their causal consequences. Then, based on our beliefs about the mechanisms of cause and effect in the world, we ask how would have things turned out differently if we had made a different decision. We use this reasoning to improve our mental models for decision-making. In contrast to typical machine learning algorithms that make decisions based exclusively on observed training data (things that *actually *happened), humans make decisions based both on observed data and imagined data (things that might have happened). Future generations of machine learning need to incorporate counterfactual reasoning if they are to reason about the world as well as humans.

After completing this section, you will be able to implement counterfactual reasoning algorithms in code. This will prepare you to implement counterfactual reasoning algorithms in automated decision-making settings in industry, such as bandits and computational advertising. You will be qualified to tackle cutting-edge problems in reinforcement learning. You will be able to evaluate machine learning algorithms for explainability and algorithmic bias.

Counterfactual deep dive

Topics

  • Counterfactual definition of causality
  • Counterfactuals vs interventions
  • Introduction to the structural causal model (SCM)
  • Multiverse counterfactuals with SCMs
  • Keystone counterfactual identities
  • Relationship between SCMs and potential outcomes

Tutorials

Programming counterfactual reasoning into AI

Dates: May 9 and 15, 2020

Topics

  • Counterfactual reasoning in bandits and reinforcement learning
  • Reparameterizing probablistic models for multiverse counterfactuals
  • Counterfactuals and intuitive physics
  • From SCMs to programs and simulations

Tutorials