user1072337
user1072337

Reputation: 12945

When running with master 'yarn' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set in the environment

I am trying to run Spark using yarn and I am running into this error:

Exception in thread "main" java.lang.Exception: When running with master 'yarn' either HADOOP_CONF_DIR or YARN_CONF_DIR must be set in the environment.

I am not sure where the "environment" is (what specific file?). I tried using:

export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop
export YARN_CONF_DIR=$HADOOP_HOME/etc/hadoop

in the bash_profile, but this doesn't seem to help.

Upvotes: 10

Views: 34572

Answers (3)

Sidharth Ramalingam
Sidharth Ramalingam

Reputation: 11

For the Windows environment, open file load-spark-env.cmd in the Spark bin folder and add the following line:

set HADOOP_CONF_DIR=%HADOOP_HOME%\etc\hadoop

Upvotes: 1

Raghavendra J.P
Raghavendra J.P

Reputation: 1

just an update to answer by Shubhangi,

 cd $SPARK_HOME/bin
 sudo nano load-spark-env.sh

add below lines , save and exit

export SPARK_LOCAL_IP="127.0.0.1"

export HADOOP_CONF_DIR="$HADOOP_HOME/etc/hadoop"

export YARN_CONF_DIR="$HADOOP_HOME/etc/hadoop"

Upvotes: -2

Shubhangi
Shubhangi

Reputation: 2264

While running spark using Yarn, you need to add following line in to spark-env.sh

export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop

Note: check $HADOOP_HOME/etc/hadoop is correct one in your environment. And spark-env.sh contains export of HADOOP_HOME as well.

Upvotes: 16

Related Questions