amine jisung
amine jisung

Reputation: 117

"Unable to instantiate SparkSession with Hive support" error when trying to process hive table with spark

I want to process hive table using spark, but when I run my program, I got this error:

Exception in thread "main" java.lang.IllegalArgumentException: Unable to instantiate SparkSession with Hive support because Hive classes are not found.

My application code

object spark_on_hive_table extends App {

  val spark = SparkSession
    .builder()
    .appName("Spark Hive Example")
    .config("spark.sql.warehouse.dir", "hdfs://localhost:54310/user/hive/warehouse")
    .enableHiveSupport()
    .getOrCreate()

  import spark.implicits._

  spark.sql("select * from pbSales").show()

}

build.sbt

version := "0.1"

scalaVersion := "2.11.12"

libraryDependencies ++= Seq(
  "org.apache.spark" %% "spark-core" % "2.3.2",
  "org.apache.spark" %% "spark-sql" % "2.3.2",
  "org.apache.spark" %% "spark-streaming" % "2.3.2",
  "org.apache.spark" %% "spark-hive" % "2.3.2" % "provided"
)

Upvotes: 2

Views: 1106

Answers (1)

mvasyliv
mvasyliv

Reputation: 1214

You should remove provided for your spark-hive dependency:

"org.apache.spark" %% "spark-hive" % "2.3.2" % "provided" 

change to

"org.apache.spark" %% "spark-hive" % "2.3.2"

Upvotes: 4

Related Questions