Re: Recreate Destroyed OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Tim, Eugen,

So what would a spec file look like for a single OSD that uses a specific
HDD (/dev/sdi) and with WAL/DB on an LV that's 25% of a specific NVMe
drive?  Regarding the NVMe, there are 3 other OSDs already using 25% each
of this NVMe for WAL/DB, but I have removed the LV that was used by the
failed OSD.  Do I need to pre-create the LV, or will 'ceph orch' do that
for me?

Thanks.

-Dave

--
Dave Hall
Binghamton University
kdhall@xxxxxxxxxxxxxx

On Thu, Oct 31, 2024 at 3:52 PM Tim Holloway <timh@xxxxxxxxxxxxx> wrote:

> I migrated from gluster when I found out it's going unsupported shortly.
> I'm really not big enough for Ceph proper, but there were only so many
> supported distributed filesystems with triple redundancy.
>
> Where I got into trouble was that I started off with Octopus and Octopus
> had some teething pains. Like stalling scheduled operations until the
> system was clan but the only way to get a clean system was to run the
> stalled operations. Pacific cured that for me.
>
> But the docs were and remain somewhat fractured between legacy and
> managed services and I managed to get into a real mess there, especially
> since I was wildly trying anything to get those stalled fixes to take.
>
> Since then, I've pretty much redefined all my OSDs with fewer but larger
> datastores and made them all managed. Now if I could just persuade the
> auto-tuner to fix the PG sizes,
>
> I'm in the process of opening a ticket account right now. The fun part
> of this is that realistically, older docs need a re-write just as much
> as the docs for the current release.
>
>     Tim
>
> On 10/31/24 15:39, Eugen Block wrote:
> > I completely understand your point of view. Our own main cluster is
> > also a bit "wild" in its OSD layout, that's why its OSDs are
> > "unmanaged" as well. When we adopted it via cephadm, I started to
> > create suitable osd specs for all those hosts and OSDs and I gave up.
> > :-D But since we sometimes also tend to experiment a bit, I rather
> > have full control over it. That's why we also have
> > osd_crush_initial_weight = 0, to check the OSD creation before letting
> > Ceph remap any PGs.
> >
> > It definitely couldn't hurt to clarify the docs, you can always report
> > on tracker.ceph.com if you have any improvement ideas.
> >
> > Zitat von Tim Holloway <timh@xxxxxxxxxxxxx>:
> >
> >> I have been slowly migrating towards spec files as I prefer
> >> declarative management as a rule.
> >>
> >> However, I think that we may have a dichotomy in the user base.
> >>
> >> On the one hand, users with dozens/hundreds of server/drives of
> >> basically identical character.
> >>
> >> On the other, I'm one who's running fewer servers and for historical
> >> reasons they tend to be wildly individualistic and often have blocks
> >> of future-use space reserved for non-ceph storage.
> >>
> >> Ceph, left to its own devices (no pun intended) can be quite
> >> enthusiastic about adopting any storage it can find. And that's great
> >> for users in the first category. Which is what the spec information
> >> in the supplied links is emphasizing. But for us lesser creatures who
> >> feel the need to manually control where each OSD and how it's
> >> configured, it's not so simple. I'm fairly certain that there's
> >> documentation on the spec file setup for that sort of stuff in the
> >> online docs, but it's located somewhere else and I cannot recall where.
> >>
> >> At any rate I would consider it very important that the different
> >> ways to setup an OSD should explicitly indicate which type of OSD
> >> will be generated in their documentation.
> >>
> >>    Tim
> >>
> >>
> >> On 10/31/24 14:28, Eugen Block wrote:
> >>> Hi,
> >>>
> >>> the preferred method to deploy OSDs in cephadm managed clusters are
> >>> spec files, see this part of the docs [0] for more information. I
> >>> would just not use the '--all-available-devices' flag, except in
> >>> test clusters, or if you're really sure that this is what you want.
> >>>
> >>> If you use 'ceph orch daemon add osd ...', you'll end up with one
> >>> (or more) OSD(s), but they will be unmanaged, as you already noted
> >>> in your own cluster. There are a couple of examples with advanced
> >>> specs (e. g. DB/WAL on dedicated devices) in the docs as well [1].
> >>> So my recommendation would be to have a suiting spec file for your
> >>> disk layout. You can always check with the '--dry-run' flag before
> >>> actually applying it:
> >>>
> >>> ceph orch apply -i osd-spec.yaml --dry-run
> >>>
> >>> Regards,
> >>> Eugen
> >>>
> >>> [0] https://docs.ceph.com/en/latest/cephadm/services/osd/#deploy-osds
> >>> [1]
> >>>
> https://docs.ceph.com/en/latest/cephadm/services/osd/#advanced-osd-service-specifications
> >>>
> >>> Zitat von Tim Holloway <timh@xxxxxxxxxxxxx>:
> >>>
> >>>> As I understand it, the manual OSD setup is only for legacy
> >>>> (non-container) OSDs. Directory locations are wrong for managed
> >>>> (containerized) OSDs, for one.
> >>>>
> >>>> Actually, the whole manual setup docs ought to be moved out of the
> >>>> mainline documentation. In their present arrangement, they make
> >>>> legacy setup sound like the preferred method. And have you noticed
> >>>> that there is no corresponding well-marked section titled
> >>>> "Authomated (cephadmin) setup?".
> >>>>
> >>>> This is how we end up with OSDs that are simultaneously legacy AND
> >>>> administered for the same OSD, since at last count there are no
> >>>> interlocks within Ceph to prevent such a mess.
> >>>>
> >>>>    Tim
> >>>>
> >>>> On 10/31/24 13:39, Dave Hall wrote:
> >>>>> Hello.
> >>>>>
> >>>>> Sorry if it appears that I am reposting the same issue under a
> >>>>> different
> >>>>> topic.  However, I feel that the problem has moved and I now have
> >>>>> different
> >>>>> questions.
> >>>>>
> >>>>> At this point I have, I believe, removed all traces of OSD.12 from my
> >>>>> cluster - based on steps in the Reef docs at
> >>>>> https://docs.ceph.com/en/reef/rados/operations/add-or-rm-osds/#.
> >>>>> I have
> >>>>> further located and removed the WAL/DB LV on an associated NVMe drive
> >>>>> (shared with 3 other OSDs).
> >>>>>
> >>>>> I don't believe the instructions for replacing an OSD (ceph-volume
> >>>>> lvm
> >>>>> prepare) still apply, so I have been trying to work with the
> >>>>> instructions
> >>>>> under ADDING AN OSD (MANUAL).
> >>>>>
> >>>>> However, since my installation is containerized (Podman), it is
> >>>>> unclear
> >>>>> which steps should be issued on the host and which within 'cephadm
> >>>>> shell'.
> >>>>>
> >>>>> There is also another ambiguity:  In step 3 the instruction is to
> >>>>> 'mkfs -t
> >>>>> {fstype}' and then to 'mount -o user_xattr'.  However, which fs type?
> >>>>>
> >>>>> After this, in step 4, the 'ceph-osd -i {osd-id} --mkfs --mkkey' gets
> >>>>> throws errors about the keyring file.
> >>>>>
> >>>>> So, are these the right instructions to be using in a containerized
> >>>>> installation?  Are there, in general, alternate documents for
> >>>>> containerized
> >>>>> installations?
> >>>>>
> >>>>> Lastly, the above cited instructions don't say anything about the
> >>>>> separate
> >>>>> WAL/DB LV.
> >>>>>
> >>>>> Please advise.
> >>>>>
> >>>>> Thanks.
> >>>>>
> >>>>> -Dave
> >>>>>
> >>>>> --
> >>>>> Dave Hall
> >>>>> Binghamton University
> >>>>> kdhall@xxxxxxxxxxxxxx
> >>>>> _______________________________________________
> >>>>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>>> _______________________________________________
> >>>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>>
> >>>
> >>> _______________________________________________
> >>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@xxxxxxx
> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux