Partial least squares (PLS)
The supervised cousin of PCR: same compress-then-regress pipeline, but the components are chosen to maximize instead of , so guides the directions. The prof’s editorial verdict, slide-quoted: “PLS often performs no better than ridge regression or PCR but it’s Swedish, so it’s like they’re meatballs, they’re not better but they sound good.” - L15-modelsel-4. Atom is correspondingly compact: know the algorithm, the contrast with PCR, and the verdict.
Definition (prof’s framing)
“It’s the same idea as the principal component analysis, only now you’re finding the principal components not as the directions of maximal variance of , but the maximal covariance of and .” - L15-modelsel-4
“PLS works similar to PCR. … But it uses the response in order to identify new features, attempts to find directions that help explain both the response and the predictors.”, slide deck (selection_regularization_presentation_lecture2.md)
Notation & setup
- Same skeleton as PCR: build orthogonal directions as linear combinations of standardized , then regress on the first .
- Difference from PCR: each is chosen to maximize (subject to ), not .
- chosen by cross-validation.
- Standardize (and typically center ).
The algorithm
To get the first PLS component:
- For each , regress on alone → simple-regression coefficient . This is proportional to .
- Set . So weights each by how well it (alone) explains .
Slide-flagged version:
” is the coefficient from the simple linear regression of onto . This coefficient is proportional to the correlation between and . PLS puts highest weight on the variables that are most strongly related to the response.”, slide deck
To get and onward:
“We regress each variable on and take the residuals. The residuals are remained info not explained by . We the[n] compute using this orthogonalized data, similarly to . We can repeat this iteration process times to get .”, slide deck
So the new direction at each step is the maximum-covariance direction in the part of -space not yet explained by previous components. Same Gram-Schmidt-style deflation idea as PCA’s “next orthogonal direction with maximum variance,” just substituting covariance for variance.
Insights & mental models
PLS is supervised; PCR is unsupervised, the only conceptual difference
The whole point. PCA / PCR picks directions of high -variance, ignoring . PLS picks directions of high , using at every step. So PLS partially fixes PCR’s biggest weakness, the no-guarantee-the-high-variance-directions-of-X-relate-to-Y problem flagged in principal-component-regression.
“It can sound like cheating, ‘it kind of sounds like you’re doing regression twice’, but it works.” - L15-modelsel-4
Verdict: not a clear winner
“In practice, PLS often performs no better than ridge regression or PCR. Supervised dimension reduction of PLS can reduce bias. It also has the potential to increase variance.”, slide deck
“All of these tend to behave similarly. A lot of things come down to looking like ridge regression, surprisingly. And conveniently, ridge regression behaves smoothly instead of this kind of discrete thing which PCR and PLS both have.” - L15-modelsel-4
The prof’s editorialized ranking, verbatim (L15-modelsel-4):
“Lasso falls somewhere between ridge and best subset regression and has some nice properties of each. For me, definitely the most interesting things are lasso and also implicit lasso. Many different ways you can get lasso and ridge regression. Those are, I think, the most interesting part of this module. But being exposed to the ideas of partial least squares regression and PCR are good because they’re fairly common, certainly PCR is very common.”
So: he likes lasso (and ridge); PCR is “fairly common” so worth knowing; PLS is niche-but-good-to-recognize.
Local color (not testable, but characterizes the prof’s framing)
“Developed in the 70s by Herman Wold (Swedish) for chemometrics, they had lots of variables, needed fast linear methods, computers were terrible, and PCA wasn’t doing it for them.” - L15-modelsel-4
“Locally relevant because… commonly used in this field called chemometrics, and one guy who’s fairly prominent in chemometrics and also in the development of these methods is actually Harold Martens, if anyone’s heard of him. He was big in this partial least squares stuff. He wrote some of the early papers in the 70s.” - L14-modelsel-3
Per scope §“Module 06”: “Detailed PLS history and chemometrics-specific tuning” is out of scope. Local color only.
PLS, PCR, ridge, converging behaviour
Slide summary:
“PLS, PCR and ridge regression tend to behave similarly. Ridge regression may be preferred because it shrinks smoothly, rather than in discrete steps.”
In summary, all three are non-sparse predictive smoothers. Use lasso or subset-selection when you actually need variable selection. Use ridge / PCR / PLS when you just want predictive accuracy and don’t care about which raw variables drive it.
Exam signals
“PLS often performs no better than ridge regression or PCR but it’s Swedish.” - L15-modelsel-4 (verbatim slide quote, the prof read this aloud)
“It’s the same idea as the principal component analysis, only now you’re finding the principal components not as the directions of maximal variance of , but the maximal covariance of and .” - L15-modelsel-4
The PLS-vs-PCR contrast (supervised vs unsupervised) is the most likely exam-handle. The prof’s verdict tells you not to over-invest: know what PLS is and how it differs from PCR; don’t memorize tuning details.
Per scope: “Detailed PLS history and chemometrics-specific tuning”, explicitly out. Concept and one-line distinction from PCR, in.
Pitfalls
- Confusing PLS with PCR. Same compress-then-regress skeleton; the only difference is what’s being maximized (covariance vs variance) and that PLS uses .
- Believing “supervised → strictly better than unsupervised.” PLS can reduce bias and increase variance, net effect is often a wash vs ridge / PCR (slide-flagged).
- Forgetting to standardize. Same as PCA / PCR.
- Treating PLS as a variable-selection method. It isn’t, like PCR, all back-transformed are typically nonzero.
- Memorizing the deflation algorithm in detail. Out of scope; the high-level “regress each on for , deflate, repeat” is enough.
Scope vs ISLP
- In scope: the PLS pipeline (same as PCR but supervised); the contrast with PCR (uses → maximizes instead of ); the algorithm at the level “regress each on to get , then deflate”; the prof’s verdict (“often no better than ridge or PCR, but supervised”).
- Look up in ISLP: §6.3.2 (pp. 286–288). Brief, ISLP doesn’t dwell on PLS either.
- Skip in ISLP / out of scope: detailed deflation algebra; chemometrics tuning; the various PLS variants (PLS-1, PLS-2, kernel PLS); history beyond “Wold, 1970s, chemometrics.” Per L15-modelsel-4 and scope.
Exercise instances
- Exercise6.9, Apply PLS on the Credit dataset and compare with previous methods (OLS / subset-selection / ridge-regression / lasso / PCR).
(Slide labels this “Recommended exercise 12”; it maps to RecEx6 problem 9 in the actual exercise file.)
How it might appear on the exam
- PLS vs PCR conceptual question: “What is the key difference between PLS and PCR?” → PLS is supervised, uses to choose components by maximizing , while PCR is unsupervised, uses only to maximize .
- Multiple choice / true-false: “PLS uses the response variable when constructing components.” → True. “PLS guarantees lower test MSE than PCR.” → False (slide-flagged: often no better).
- Method comparison from a results table: PLS / PCR / ridge / lasso side by side on the same data. Recognize that the first three behave similarly; lasso is the differentiator if sparsity matters.
- Probably not asked: detailed PLS algorithm pseudocode; chemometrics applications; history.
Related
- principal-component-regression: the unsupervised cousin; PLS is “PCR but uses .”
- principal-component-analysis: PCA’s variance-maximization is what PLS replaces with covariance-maximization.
- ridge-regression: slide-flagged behaviour: PLS, PCR, ridge “tend to behave similarly,” ridge preferred for smoothness.
- lasso: the variable-selecting alternative; different philosophy entirely.
- dimensionality-reduction: the umbrella concept.
- standardization: required preprocessing.
- cross-validation: for choosing .