Reputation: 479
I tried to use Hadoop, then installed and could use as a stand-alone mode. But when I use as a psuedo-distributed mode, below message was occurred and didn't proceed the process.
17/10/24 02:04:15 INFO client.RMProxy: Connecting to ResourceManager at localhost/127.0.0.1:8032
17/10/24 02:04:16 INFO input.FileInputFormat: Total input files to process : 10
17/10/24 02:04:16 INFO mapreduce.JobSubmitter: number of splits:10
17/10/24 02:04:17 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1508778206216_0001
17/10/24 02:04:17 INFO impl.YarnClientImpl: Submitted application application_1508778206216_0001
17/10/24 02:04:17 INFO mapreduce.Job: The url to track the job: http://MacBook.local:8088/proxy/application_1508778206216_0001/
17/10/24 02:04:17 INFO mapreduce.Job: Running job: job_1508778206216_0001
I checked localhost:50070 and there are a working Datanode. I show my setup procedure.
①Install Hadoop
brew install hadoop
②hadoop configurations ○libexec/etc/hadoop/core-site.xml
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost:9000</value>
</property>
</configuration>
○libexec/etc/hadoop/hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
○libexec/etc/hadoop/yarn-site.xml
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
</configuration>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>localhost</value>
</property>
○libexec/etc/hadoop/mapred-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapred.job.tracker</name>
<value>localhost:54311</value>
</property>
</configuration>
③start hadoop
sbin/start-all.sh
with jps
worked.
④Run of Hadoop
hadoop jar libexec/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.8.1.jar pi 10 100000
then
Number of Maps = 10
Samples per Map = 100000
17/10/24 02:04:12 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Wrote input for Map #0
Wrote input for Map #1
Wrote input for Map #2
Wrote input for Map #3
Wrote input for Map #4
Wrote input for Map #5
Wrote input for Map #6
Wrote input for Map #7
Wrote input for Map #8
Wrote input for Map #9
Starting Job
17/10/24 02:04:15 INFO client.RMProxy: Connecting to ResourceManager at localhost/127.0.0.1:8032
17/10/24 02:04:16 INFO input.FileInputFormat: Total input files to process : 10
17/10/24 02:04:16 INFO mapreduce.JobSubmitter: number of splits:10
17/10/24 02:04:17 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1508778206216_0001
17/10/24 02:04:17 INFO impl.YarnClientImpl: Submitted application application_1508778206216_0001
17/10/24 02:04:17 INFO mapreduce.Job: The url to track the job: http://MacBook.local:8088/proxy/application_1508778206216_0001/
17/10/24 02:04:17 INFO mapreduce.Job: Running job: job_1508778206216_0001
The process never proceeded. Please tell me the reason why it doesn't work. And I accessed "http://macbook.local:8088/proxy/application_1508759907777_0001/" to check Jobtracker, but an error code "ERR_EMPTY_RESPONSE" was returned.
Upvotes: 0
Views: 93
Reputation: 133
First, note that Hadoop 2.x uses the YARN Resource Managers and Node Managers in place of Job Trackers and Task Trackers.
Instead, you can try adding the following property to yarn-site.xml
:
<property>
<name>yarn.resourcemanager.hostname</name>
<value>localhost</value>
</property>
I'm not sure what is supposed to happen to jobtracker
properties in Hadoop 2.x, but perhaps it is interfering. Dropping it and explicitly setting the resourcemanager.hostname
may resolve.
You can test if it's available by opening localhost:8032
in a browser. More information on setting up a pseudo-distributed cluster is available in the Apache Hadoop docs.
Upvotes: 0