Getting Started with Apache Spark
Apache Spark is a powerful engine for large-scale data processing. Scala is its native language.
Setting Up Spark with sbt
libraryDependencies += "org.apache.spark" %% "spark-core" % "3.4.0"
libraryDependencies += "org.apache.spark" %% "spark-sql" % "3.4.0"
Starting a SparkSession
import org.apache.spark.sql.SparkSession
val spark = SparkSession.builder()
.appName("MyApp")
.master("local[*]")
.getOrCreate()
This initializes Spark for local execution.