Reputation: 4407
Most of cases I encountered were where data was being migrated in/out of HDFS. And I am entirely new to Hadoop!
But my requirement is to move large binary files from various sources in a single JCR compliant flat file repository. my question is
1) Is Hadoop a good choice for Read from X and write to Y (Both non HDFS systems) scenario?
2) Will any other custom framework or something like Spring batch will serve same purpose? Meaning it is just a problem of IO and network?
Upvotes: 1
Views: 941
Reputation: 2724
Yes you can use hadoop, if your X and Y can handle multiple connections and will give better performance if you retrieve and store data in parallel.
There is a framework called sqoop that could meet your needs better.
Upvotes: 1