Duc Duong
Duc Duong

Reputation: 21

Hadoop is configured to use IP address but still connect to localhost

I have configured Hadoop to use IP address instead of localhost but when I started it, it still posted an error that can not ssh to localhost.

hdfs-site.xml:

<configuration>
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///<hdfs_path>/hdfs/datanode</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///<hdfs_path>/hdfs/namenode</value>
</property>
</configuration>

core-site.xml:

<name>fs.defaultFS</name>
<value>hdfs://<IP_address>/</value>  
<description>NameNode URI</description>

When I started Hadoop, the below error appeared:

localhost: ssh_exchange_identification: Connection closed by remote host

When I tried to allow shh to localhost in our test server, the error disappeared. But problem is that in our production server, ssh to localhost is not allowed. I tried to avoid using localhost and use IP address instead, but why Hadoop still needs to ssh to localhost?

Upvotes: 0

Views: 819

Answers (1)

OneCricketeer
OneCricketeer

Reputation: 191874

Only the start-dfs, start-yarn, or start-all shell scripts use SSH, and the SSH connection has nothing to do with the XML files.

That's not required; you can run hadoop namenode and hadoop datanode, and the YARN daemon processes directly. However, you'll still need to somehow get into a shell of each machine to run those commands, if they don't start at boot or ever fail to (re)start

Upvotes: 1

Related Questions