Re: ceph progress bar stuck and 3rd manager not deploying

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



No problem, and it looks like they will. Glad it worked out for you!

David

On Thu, Sep 9, 2021 at 9:31 AM mabi <mabi@xxxxxxxxxxxxx> wrote:
>
> Thank you Eugen. Indeed the answer went to Spam :(
>
> So thanks to David for his workaround, it worked like a charm. Hopefully these patches can make it into the next pacific release.
>
> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
>
> On Thursday, September 9th, 2021 at 2:33 PM, Eugen Block <eblock@xxxxxx> wrote:
>
> > You must have missed the response to your thread, I suppose:
> >
> > https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/YA5KLI5MFJRKVQBKUBG7PJG4RFYLBZFA/
> >
> > Zitat von mabi mabi@xxxxxxxxxxxxx:
> >
> > > Hello,
> > >
> > > A few days later the ceph status progress bar is still stuck and the
> > >
> > > third mon is for some unknown reason still not deploying itself as
> > >
> > > can be seen from the "ceph orch ls" output below:
> > >
> > > ceph orch ls
> > >
> > > NAME PORTS RUNNING REFRESHED AGE PLACEMENT
> > >
> > > alertmanager ?:9093,9094 1/1 3m ago 5w count:1
> > >
> > > crash 7/7 3m ago 5w *
> > >
> > > grafana ?:3000 1/1 3m ago 5w count:1
> > >
> > > mgr 2/2 3m ago 4w count:2;label:mgr
> > >
> > > mon 2/3 3m ago 16h count:3;label:mon
> > >
> > > node-exporter ?:9100 7/7 3m ago 5w *
> > >
> > > osd 1/1 3m ago - <unmanaged>
> > >
> > > prometheus ?:9095 1/1 3m ago 5w count:1
> > >
> > > Is this a bug in cephadm? and is there a workaround?
> > >
> > > Thanks for any hints.
> > >
> > > ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
> > >
> > > On Tuesday, September 7th, 2021 at 2:30 PM, mabi mabi@xxxxxxxxxxxxx wrote:
> > >
> > > > Hello
> > > >
> > > > I have a test ceph octopus 16.2.5 cluster with cephadm out of 7
> > > >
> > > > nodes on Ubuntu 20.04 LTS bare metal. I just upgraded each node's
> > > >
> > > > kernel and performed a rolling reboot and now the ceph -s output is
> > > >
> > > > stuck somehow and the manager service is only deployed to two nodes
> > > >
> > > > instead of 3 nodes. Here would be the ceph -s output:
> > > >
> > > > cluster:
> > > >
> > > > id: fb48d256-f43d-11eb-9f74-7fd39d4b232a
> > > >
> > > > health: HEALTH_WARN
> > > >
> > > > OSD count 1 < osd_pool_default_size 3
> > > >
> > > > services:
> > > >
> > > > mon: 2 daemons, quorum ceph1a,ceph1c (age 25m)
> > > >
> > > > mgr: ceph1a.guidwn(active, since 25m), standbys: ceph1c.bttxuu
> > > >
> > > > osd: 1 osds: 1 up (since 30m), 1 in (since 3w)
> > > >
> > > > data:
> > > >
> > > > pools: 0 pools, 0 pgs
> > > >
> > > > objects: 0 objects, 0 B
> > > >
> > > > usage: 5.3 MiB used, 7.0 TiB / 7.0 TiB avail
> > > >
> > > > pgs:
> > > >
> > > > progress:
> > > >
> > > > Updating crash deployment (-1 -> 6) (0s)
> > > >
> > > >       [............................]
> > > >
> > > >
> > > > Ignore the HEALTH_WARN with of the OSD count because I have not
> > > >
> > > > finished to deploy all 3 OSDs. But you can see that the progress
> > > >
> > > > bar is stuck and I have only 2 managers, the third manager does not
> > > >
> > > > seem to start as can be seen here:
> > > >
> > > > $ ceph orch ps|grep stopped
> > > >
> > > > mon.ceph1b ceph1b stopped 4m ago 4w - 2048M <unknown> <unknown> <unknown>
> > > >
> > > > It looks like the orchestrator is stuck and does not continue it's
> > > >
> > > > job. Any idea how I can get it unstuck?
> > > >
> > > > Best regards,
> > > >
> > > > Mabi
> > >
> > > ceph-users mailing list -- ceph-users@xxxxxxx
> > >
> > > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> > ceph-users mailing list -- ceph-users@xxxxxxx
> >
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux