Getting Started with Apache Spark

Apache Spark is a powerful engine for large-scale data processing. Scala is its native language.

Setting Up Spark with sbt

libraryDependencies += "org.apache.spark" %% "spark-core" % "3.4.0"
libraryDependencies += "org.apache.spark" %% "spark-sql" % "3.4.0"

Starting a SparkSession

import org.apache.spark.sql.SparkSession

val spark = SparkSession.builder()
  .appName("MyApp")
  .master("local[*]")
  .getOrCreate()

This initializes Spark for local execution.

← PrevNext →