- PySpark Cookbook
- Denny Lee Tomasz Drabas
- 79字
- 2021-06-18 19:06:36
Reading data from files
For this recipe, we will create an RDD by reading a local file in PySpark. To create RDDs in Apache Spark, you will need to first install Spark as noted in the previous chapter. You can use the PySpark shell and/or Jupyter notebook to run these code samples. Note that while this recipe is specific to reading local files, a similar syntax can be applied for Hadoop, AWS S3, Azure WASBs, and/or Google Cloud Storage:

推薦閱讀
- iOS面試一戰到底
- Node.js 10實戰
- 少年輕松趣編程:用Scratch創作自己的小游戲
- 動手玩轉Scratch3.0編程:人工智能科創教育指南
- Scala Design Patterns
- Java加密與解密的藝術(第2版)
- Mastering Android Development with Kotlin
- C++從入門到精通(第5版)
- Mastering Xamarin.Forms(Second Edition)
- Internet of Things with ESP8266
- Flink技術內幕:架構設計與實現原理
- Python預測之美:數據分析與算法實戰(雙色)
- Learning Cocos2d-JS Game Development
- MongoDB Cookbook
- Android從入門到精通