A linear algebra book I’m working on, and will post online.

The book employs a novel approach to teaching linear algebra by centering on singular value decomposition (SVD). This vantage point provides a deeper understanding of algorithms, like PCA, quadratic programming, and regressions, and allows the reader to develop their intuition for how to modify and apply these algorithms in real world applications.

The book interweaves theory and application, beginning with SVD, from which PCA and collaborative filtering immediately follow. Regressions can be understood via projection matrices (matrices with singular values all equal to 1) or via pseudo-inverses (inverses on the non-zero singular values). Quadratic programming can be understood via positive definite matrices (matrices with an SVD having inverse isometries and all positive singular values). Recognizing that most applied linear algebra follows from a solid understanding of SVD makes it easier to understand these hard topics.

After establishing a solid theoretical basis, the book dives into application. It discusses which model to use when, how to interpret model outputs, and what modeling choices are available for each algorithm. It grounds this discussion in concrete applications, like the Netflix Prize for collaborative filtering. Further, the book provides real-world exercises, such as, “You and a coworker are building an insurance model on credit data. The model will be a logistic regression to predict the probability of an insurance claim. You have historical claim data. You think you should use only PLS, and your coworker thinks you should only use PCA. How can you decide with data which model is better?”

MathPCASVDMatricesCollaborative FilteringQuadratic Programming Link