Welcome to some practical explanations to Apache Spark with Scala. There is even **_Python supported Spark is available which is PySpark. _**For the sake of this post, I am continuing with Scala with my windows Apache Spark installation.
Initialization of Spark Shell
spark-shell_ command provides a simple way to learn the Spark API, as well as a powerful tool to analyze data interactively. _It is available in either Scala or Python
Let’s discuss some terms logged for spark-shell command.
SparkContext
(sc) is the entry point for Spark functionality. A Spark Context represents the connection to a Spark cluster and can be used to create RDDs in the cluster. Only one SparkContext should be active per JVM. All about Sparck Context constructors and Methods can be found here official link**.**SparkSession
(spark) is the entry to programming Spark with the Dataset and DataFrame API. It is one of the very first objects you create while developing a Spark SQL applicationSpark-Shell Commands:
#scala #big-data #apache-spark #apache-spark-rdd