- Mastering Apache Spark 2.x(Second Edition)
- Romeo Kienzler
- 110字
- 2021-07-02 18:55:29
Using SQL
After using the previous Scala example to create a data frame from a JSON input file on HDFS, we can now define a temporary table based on the data frame and run SQL against it.
The following example shows you the temporary table called washing_flat being defined and a row count being created using count(*):

The schema for this data was created on the fly (inferred). This is a very nice function of the Apache Spark DataSource API that has been used when reading the JSON file from HDFS using the SparkSession object. However, if you want to specify the schema on your own, you can do so.
推薦閱讀
- Visual Basic .NET程序設計(第3版)
- LabVIEW 2018 虛擬儀器程序設計
- 零基礎PHP學習筆記
- 劍指Offer(專項突破版):數據結構與算法名企面試題精講
- HBase從入門到實戰
- Java 9 Programming Blueprints
- JavaScript 程序設計案例教程
- 程序是怎樣跑起來的(第3版)
- MATLAB for Machine Learning
- Python大學實用教程
- Unity 2017 Game AI Programming(Third Edition)
- 多媒體技術及應用
- Google Adsense優化實戰
- Head First Kotlin程序設計
- Python數據可視化之matplotlib實踐