- Frank Kane's Taming Big Data with Apache Spark and Python
- Frank Kane
- 91字
- 2021-07-02 21:12:19
What is Spark?
According to Apache, Spark is a fast and general engine for large-scale data processing. This is actually a really good summary of what it's all about. If you have a really massive dataset that can represent anything - weblogs, genomics data, you name it - Spark can slice and dice that data up. It can distribute the processing among a huge cluster of computers, taking a data analysis problem that's just too big to run on one machine and divide and conquer it by splitting it up among multiple machines.
推薦閱讀
- 黑客攻防從入門到精通(實戰秘笈版)
- Expert C++
- 深入理解Bootstrap
- Getting started with Google Guava
- JIRA 7 Administration Cookbook(Second Edition)
- Android Development with Kotlin
- EPLAN實戰設計
- Unity 5.x By Example
- Windows Phone 7.5:Building Location-aware Applications
- OpenMP核心技術指南
- 計算機應用基礎項目化教程
- CryENGINE Game Programming with C++,C#,and Lua
- C++ Windows Programming
- MATLAB語言及編程實踐:生物數學模型應用
- LabVIEW案例實戰