forked from svpino/ml.school
-
Notifications
You must be signed in to change notification settings - Fork 0
/
index.qmd
86 lines (71 loc) · 4.19 KB
/
index.qmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---
title: "Introduction"
listing:
contents: posts
sort: "date desc"
type: default
categories: false
---
Welcome to the program!
## Program Structure
#### Session 1 - Production Machine Learning Is Different
* An overview of the components of a machine learning system
* The role of data in real-world applications
* 4 popular sampling strategies to collect data
* Labeling strategies and a quick introduction to weak supervision and active learning
* Data splitting and how data leakage can destroy your models
* Building good features and a quick introduction to imputation, standardization, and encoding
* Processing data at scale using data parallelism
* Using pipelines to orchestrate machine learning workflows
* A template architecture of a production-ready machine learning system
* Understanding SageMaker’s Processing Step and Processing Jobs
#### Session 2 - Building Models And The Training Pipeline
* The first rule of Machine Learning Engineering
* A 3-step process to solve a problem using machine learning
* 9 tips to select the best machine learning model for your solution
* Strategies for working with imbalanced data, dealing with rare events, and a quick introduction to cost-sensitive learning
* The reason you should not balance your data
* An introduction to hyperparameter tuning
* The importance of reproducibility and a quick introduction to experiment tracking
* Distributed Training using data and model parallelism
* Understanding SageMaker’s Training and Tuning Steps, and Training and Tuning Jobs
#### Session 3 - Evaluating and Versioning Models
* The difference between good models and useful models
* Framing evaluation metrics in the context of business performance
* An 8-step process to evaluate machine learning models
* Introduction to backtesting
* How to deal with disproportionally important examples and rare cases
* Strategies to determine whether a model is fair and robust to future changes
* A 3-step process to perform error analysis and measure the impact of potential improvements
* How to determine whether individual predictions are useful
* Evaluating Large Language Models using Supervised Evaluation and Self-Evaluation
* An introduction to model versioning
* Understanding SageMaker’s Model Registry, Condition, and Model Steps
#### Session 4 - Deploying Models and Serving Predictions
* How do model performance, speed, and cost affect models in production
* Latency, throughput, and their relationships
* Understanding on-demand inference and batch inference and when to use each one
* How to make models run fast using model compression and a quick introduction to quantization and knowledge distillation
* Deploying models in dedicated and multi-model endpoints
* A comparison of the tools you can use to serve predictions
* Designing a 3-component inference pipeline
* Understanding the internal structure of a SageMaker Endpoint
* Understanding SageMaker's PipelineModel and Amazon EventBridge
#### Session 5 - Data Distribution Shifts And Model Monitoring
* The 3 most common problems your model will face in production
* An introduction to data distribution shifts, edge cases, and unintended feedback loops
* Catastrophic predictions and the problem with edge cases
* Understanding covariate shift and concept drift
* Monitoring schema violations, data statistics, model performance, prediction distribution, and changes in user feedback
* The 3 strategies to keep your models working despite data distribution shifts
* Understanding SageMaker’s Transform Step, QualityCheck Step, Transform Jobs, and Monitoring Jobs
#### Session 6 - Continual Learning And Testing in Production
* The importance of Continual Learning and why every company wants to to do it
* 3 challenges when implementing Continual Learning
* A 4-step plan to implement Continual Learning
* How to determine what data to use to retrain a model
* A 3-step progressive plan to decide how frequently you should retrain your models
* The differences between training from scratch and incremental training
* An introduction to Testing in Production
* 5 strategies to test models in production: A/B testing, shadow deployments, canary releases, interleaving experiments, and multi-armed bandits
* Highlights from the program