Discussion:
[QUESTION] Interworking 2 seperate Hadoop instances
Dave Morris
2017-02-16 07:15:32 UTC
Permalink
Hi everyone,

I am confronted with quite a difficult problem, how to connect 2 separate
Hadoop instances, each having their own HDFS's, HBASE's etc.

My first though would be what if they share a Master node, surely this will
allow any Mapreduce or Spark request to be run against both instances?
Comments?

If the above idea is feasible, how would one ensure that the data
collection remain in each respective instance, and not be distributed among
both instances?

Kind Regards
Dave
Daniel Templeton
2017-02-16 16:13:54 UTC
Permalink
If they share the same master node, then they're really only one
cluster, so everything will work as expected. If you want the two
clusters to remain independent, what you want is federation. I believe
HDFS already supports federation, and YARN federation work is underway.

Daniel
Post by Dave Morris
Hi everyone,
I am confronted with quite a difficult problem, how to connect 2 separate
Hadoop instances, each having their own HDFS's, HBASE's etc.
My first though would be what if they share a Master node, surely this will
allow any Mapreduce or Spark request to be run against both instances?
Comments?
If the above idea is feasible, how would one ensure that the data
collection remain in each respective instance, and not be distributed among
both instances?
Kind Regards
Dave
---------------------------------------------------------------------
To unsubscribe, e-mail: general-***@hadoop.apache.org
For additional commands, e-mail: general-***@hadoop.apache.org

Loading...