- Deep Learning By Example
- Ahmed Menshawy
- 157字
- 2021-06-24 18:52:36
Feature selection
The number of explanatory features (input variables) of a sample can be enormous wherein you get xi=(xi1, xi2, xi3, ... , xid) as a training sample (observation/example) and d is very large. An example of this can be a document classification task3, where you get 10,000 different words and the input variables will be the number of occurrences of different words.
This enormous number of input variables can be problematic and sometimes a curse because we have many input variables and few training samples to help us in the learning procedure. To avoid this curse of having an enormous number of input variables (curse of dimensionality), data scientists use dimensionality reduction techniques in order to select a subset from the input variables. For example, in the text classification task they can do the following:
- Extracting relevant inputs (for instance, mutual information measure)
- Principal component analysis (PCA)
- Grouping (cluster) similar words (this uses a similarity measure)
- 大數據導論:思維、技術與應用
- 集成架構中型系統
- 大學計算機基礎:基礎理論篇
- Mastering Proxmox(Third Edition)
- 返璞歸真:UNIX技術內幕
- Visual C# 2008開發技術實例詳解
- 中國戰略性新興產業研究與發展·增材制造
- 工業自動化技術實訓指導
- Dreamweaver+Photoshop+Flash+Fireworks網站建設與網頁設計完全實用
- 貫通開源Web圖形與報表技術全集
- Linux Shell Scripting Cookbook(Third Edition)
- 智能制造系統及關鍵使能技術
- Learning Cassandra for Administrators
- Data Analysis with R(Second Edition)
- Oracle 11g基礎與提高