Reputation: 5870
I'm using Spark1.3.1 trying to read data from HDFS as:
val sc = new SparkContext(sparkConf)
val lines = sc.textFile("hdfs://192.168.0.104:9000/cur/part-r-02554")
I met the following excepton:
Exception in thread "main" java.io.IOException: Failed on local exception:
com.google.protobuf.InvalidProtocolBufferException: Protocol message end-group
tag did not match expected tag.; Host Details : local host is:
"hadoop104/192.1168.1.104"; destination host is: "hadoop104":9000;
Upvotes: 0
Views: 734
Reputation: 727
Try looking for property fs.defaultFS
or fs.default.name
in your core-site.xml
. Check if 192.168.0.104
is configured as value
, and not a hostname
.
If a hostname is configured as a value
, this is bound to give you an error - as this is very strictly followed. Either, use whatever is configured in core-site.xml
or do not use an IP/hostname and just go ahead with hdfs:/cur/part-r-02554
Upvotes: 1