On Fri, Nov 15, 2019 at 4:39 PM Wido den Hollander <wido@xxxxxxxx> wrote: > > > > On 11/15/19 4:25 PM, Paul Emmerich wrote: > > On Fri, Nov 15, 2019 at 4:02 PM Wido den Hollander <wido@xxxxxxxx> wrote: > >> > >> I normally use LVM on top > >> of each device and create 2 LVs per OSD: > >> > >> - WAL: 1GB > >> - DB: xx GB > > > > Why? I've seen this a few times and I can't figure out what the > > advantage of doing this explicitly on the LVM level instead of relying > > on BlueStore to handle this. > > > > If the WAL+DB are on a external device you want the WAL to be there as > well. That's why I specify the WAL separate. > > This might be an 'old habbit' as well. But the WAL will be placed onto the DB device if it isn't explicitly specified, so there's no advantage to having a separate partition. Paul > > Wido > > > > > Paul > > > >> > >>> > >>> > >>> The initial cluster is +1PB and we’re planning to expand it again with > >>> 1PB in the near future to migrate our data. > >>> > >>> We’ll only use the system thru the RGW (No CephFS, nor block device), > >>> and we’ll store “a lot” of small files on it… (Millions of files a day) > >>> > >>> > >>> > >>> The reason I’m asking it, is that I’ve been able to break the test > >>> system (long story), causing OSDs to fail as they ran out of space… > >>> Expanding the disks (the block DB device as well as the main block > >>> device) failed with the ceph-bluestore-tool… > >>> > >>> > >>> > >>> Thanks for your answer! > >>> > >>> > >>> > >>> Kristof > >>> > >>> > >>> _______________________________________________ > >>> ceph-users mailing list > >>> ceph-users@xxxxxxxxxxxxxx > >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>> > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@xxxxxxxxxxxxxx > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com