Hi Zoltan, thanks for the answer.
Because replacing hdfs:// with ceph:// and use CephFs doesn't work for all haddop componentes out of the box (unless in my tests), for example I had issues with Hbase, then with Yarn, Hue, etc (I'm using the cloudera distribution but I also tried with separate
components). And besides the need to add jars and bindings to each node to get them work, there are a lot of places (xmls, configuration) where the "hdfs for ceph" replacement need to be made.
Giving this issues, I thought that mounting ceph as a local directory and then use this "virtual dirs" as the haddop dfs dirs, would be easier and will work better (less configuration problems, and only changing the dfs dirs will make all components work without any more changes).
Of course I can be totally wrong, and it's a core change to do this, that's why I think I should ask here first :)
Thanks!
De: Zoltan Arnold Nagy <zoltan@xxxxxxxxxxxxxxxxxx>
Enviado: jueves, 04 de febrero de 2016 06:07 p.m. Para: John Spray Cc: Jose M; ceph-users@xxxxxxxxxxxxxx Asunto: Re: Ceph and hadoop (fstab insted of CephFS) Might be totally wrong here, but it’s not layering them but replacing hdfs:// URLs with ceph:// URLs so all the mapreduce/spark/hbase/whatever is on top can use CephFS directly which is not a bad thing to do (if it works) :-)
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com