Book contents
- Frontmatter
- Dedication
- Contents
- Preface
- Part One Fundamentals
- 1 Introduction
- 2 Features, Combined: Normalization, Discretization and Outliers
- 3 Features, Expanded: Computable Features, Imputation and Kernels
- 4 Features, Reduced: Feature Selection, Dimensionality Reduction and Embeddings
- 5 Advanced Topics: Variable-Length Data and Automated Feature Engineering
- Part II Case Studies
- Bibliography
- Index
4 - Features, Reduced: Feature Selection, Dimensionality Reduction and Embeddings
from Part One - Fundamentals
Published online by Cambridge University Press: 29 May 2020
- Frontmatter
- Dedication
- Contents
- Preface
- Part One Fundamentals
- 1 Introduction
- 2 Features, Combined: Normalization, Discretization and Outliers
- 3 Features, Expanded: Computable Features, Imputation and Kernels
- 4 Features, Reduced: Feature Selection, Dimensionality Reduction and Embeddings
- 5 Advanced Topics: Variable-Length Data and Automated Feature Engineering
- Part II Case Studies
- Bibliography
- Index
Summary
This chapter presents a staple of Feature Engineering: the automatic reduction of features, either by direction selection or by projection to a smaller feature space.Central to Feature Engineering are efforts to reduce the number of features, as uninformative features bloat the ML model with unnecessary parameters. In turn, too many parameters then either produces suboptimal results, as they are easy to overfit, or require large amounts of training data. These efforts are either by explicitly dropping certain features (feature selection) or mapping the feature vector, if it is sparse, into a lower, denser dimension (dimensionality reduction). There are also cover some algorithms that perform feature selection as part of their inner computation (embedded feature selection or regularization). Feature selection takes the spotlight within Feature Engineering due to its intrinsic utility for Error Analysis. Some techniques such as feature ablation using wrapper methods are used as the starting step before a feature drill down. Moreover, as feature selection helps build understandable models, it intertwines with Error Analysis as the analysis profits from such understandable models.
Keywords
- Type
- Chapter
- Information
- The Art of Feature EngineeringEssentials for Machine Learning, pp. 79 - 111Publisher: Cambridge University PressPrint publication year: 2020