- Scala for Machine Learning(Second Edition)
- Patrick R. Nicolas
- 143字
- 2021-07-08 10:43:10
Chapter 5. Dimension Reduction
As described in the Assessing a model/overfitting section of Chapter 2, Data Pipelines, the indiscriminative reliance of a large number of features may cause overfitting; the model may become so tightly coupled with the training set that different validation sets will generate a vastly different outcome and quality metrics such as AuROC.
Dimension reduction techniques alleviate these problems by detecting features that have little influence on the overall model behavior.
This chapter introduces three categories of dimension reduction techniques with two implementations in Scala:
- Divergence with an implementation of the Kullback-Leibler distance
- Principal components analysis
- Estimation of low dimension feature space for nonlinear models
Other types of methodologies used to reduce the number of features such as regularization or singular value decomposition are discussed in future chapters.
But first, let's start our investigation by defining the problem.
- Cocos2D-X權(quán)威指南(第2版)
- 在最好的年紀(jì)學(xué)Python:小學(xué)生趣味編程
- Learn to Create WordPress Themes by Building 5 Projects
- Python進(jìn)階編程:編寫更高效、優(yōu)雅的Python代碼
- CouchDB and PHP Web Development Beginner’s Guide
- Spring Boot Cookbook
- Mastering Drupal 8 Views
- Hands-On Automation Testing with Java for Beginners
- Django 3.0入門與實(shí)踐
- JSP程序設(shè)計(jì)實(shí)例教程(第2版)
- 零基礎(chǔ)學(xué)C語言(第4版)
- PowerDesigner 16 從入門到精通
- C# 7.1 and .NET Core 2.0:Modern Cross-Platform Development(Third Edition)
- Clojure Web Development Essentials
- Web前端開發(fā)精品課:HTML5 Canvas開發(fā)詳解