Oscar Segarra wrote:
I'd like to hear your opinion about theese two configurations:You mean 1 OSD daemon on top of RAID5? I don't think I'd do that. You'll probably want redundancy at Ceph's level anyhow, and then where is the point...? I'm a little bit worried that 8 osd daemons can affect performance because all jobs running and scrubbing.If you ran RAID instead of Ceph, RAID might still perform better. But I don't believe anything much changes for the better if you run the Ceph on top of RAID rather than on top of individual OSD, unless your configuration is bad. I generally don't think you have to worry that much that a reasonably modern machine can't handle running a few extra jobs, either. But you could certainly do some tests on your hardware to be sure. Another question is the procedure of a replacement of a failed disk. In case of a big RAID, replacement is direct. In case of many OSDs, the procedure is a little bit tricky.I wasn't using Ceph in 2014, but at least in my limited experience, today the most important step is done when you add the new drive and activate an OSD on it. You probably still want to remove the leftovers of the old failed OSD for it to not clutter your list, but as far as I can tell replication and so on will trigger *before* you remove it. (There is a configurable timeout for how long an OSD can be down, after which the OSD is essentially treated as dead already, at which point replication and rebalancing starts). -Michael |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com