- Artificial Intelligence for Big Data
- Anand Deshpande Manish Kumar
- 342字
- 2021-06-25 21:57:12
Learning from Big Data
In the first two chapters, we set the context for intelligent machines with the big data revolution and how big data is fueling rapid advances in artificial intelligence. We also emphasized the need for a global vocabulary for universal knowledge representation. We have also seen how that need is fulfilled with the use of ontologies and how ontologies help construct a semantic view of the world.
The quest is for the knowledge, which is derived from information, which is in turn derived from the vast amounts of data that we are generating. Knowledge facilitates a rational decision-making process for machines that complements and augments human capabilities. We have seen how the Resource Description Framework (RDF) provides the schematic backbone for the knowledge assets along with Web Ontology Language (OWL) fundamentals and the query language for RDFs (SPARQL).
In this chapter, we are going to look at some of the basic concepts of machine learning and take a deep pe into some of the algorithms. We will use Spark's machine learning libraries. Spark is one of the most popular computer frameworks for the implementation of algorithms and as a generic computation engine on big data. Spark fits into the big data ecosystem well, with a simple programming interface, and very effectively leverages the power of distributed and resilient computing frameworks. Although this chapter does not assume any background with statistics and mathematics, it will greatly help if the reader has some programming background, in order to understand the code snippets and to try and experiment with the examples.
In this chapter, we will see broad categories of machine learning in supervised and unsupervised learning, before taking a deep pe, with examples, into:
- Regression analysis
- Data clustering
- K-means
- Data dimensionality reduction
- Singular value decomposition
- Principal component analysis (PCA)
In the end, we will have an overview of the Spark programming model and Spark's Machine Learning library (Spark MLlib). With all this background knowledge at our disposal, we will implement a recommendation system to conclude this chapter.
- 從零開始學Hadoop大數據分析(視頻教學版)
- iOS and OS X Network Programming Cookbook
- 數據驅動:從方法到實踐
- 一個64位操作系統的設計與實現
- 深入淺出Greenplum分布式數據庫:原理、架構和代碼分析
- 區塊鏈技術應用與實踐案例
- Hadoop 3實戰指南
- Web Services Testing with soapUI
- Access 2010數據庫程序設計實踐教程
- 實現領域驅動設計
- Unity Game Development Blueprints
- 數據應用工程:方法論與實踐
- Oracle 內核技術揭密
- AI Crash Course
- 數據產品經理寶典:大數據時代如何創造卓越產品