Re: upgrade problem nautilus 14.2.15 -> 14.2.18? (Broken ceph!)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



In each host's ceph.conf, put:

[global]
public addr = <host ip>
cluster addr = <host ip>


Or use the public network / cluster network options = a.b.c.0/24 or
however your network is defined.

-- dan


On Thu, Mar 25, 2021 at 8:50 PM Simon Oosthoek <s.oosthoek@xxxxxxxxxxxxx> wrote:
>
> On 25/03/2021 20:42, Dan van der Ster wrote:
> > netstat -anp | grep LISTEN | grep mgr
> >
> > has it bound to 127.0.0.1 ?
> >
> > (also check the other daemons).
> >
> > If so this is another case of https://tracker.ceph.com/issues/49938
>
> Do you have any idea for a workaround (or should I downgrade?). I'm
> running ceph on ubuntu 18.04 LTS
>
> this seems to be happening on the mons/mgrs and osds
>
> Cheers
>
> /Simon
> >
> > -- dan
> >
> > On Thu, Mar 25, 2021 at 8:34 PM Simon Oosthoek <s.oosthoek@xxxxxxxxxxxxx> wrote:
> >>
> >> Hi
> >>
> >> I'm in a bit of a panic :-(
> >>
> >> Recently we started attempting to configure a radosgw to our ceph
> >> cluster, which was until now only doing cephfs (and rbd wss working as
> >> well). We were messing about with ceph-ansible, as this was how we
> >> originally installed the cluster. Anyway, it installed nautilus 14.2.18
> >> on the radosgw and I though it would be good to pull up the rest of the
> >> cluster to that level as well using our tried and tested ceph upgrade
> >> script (it basically does an update of all ceph nodes one by one and
> >> checks whether ceph is ok again before doing the next)
> >>
> >> After the 3rd mon/mgr was done, all pg's were unavailable :-(
> >> obviously, the script is not continuing, but ceph is also broken now...
> >>
> >> The message deceptively is: HEALTH_WARN Reduced data availability: 5568
> >> pgs inactive
> >>
> >> That's all PGs!
> >>
> >> I tried as a desperate measure to upgrade one ceph OSD node, but that
> >> broke as well, the osd service on that node gets an interrupt from the
> >> kernel....
> >>
> >> the versions are now like:
> >> 20:29 [root@cephmon1 ~]# ceph versions
> >> {
> >>       "mon": {
> >>           "ceph version 14.2.18
> >> (befbc92f3c11eedd8626487211d200c0b44786d9) nautilus (stable)": 3
> >>       },
> >>       "mgr": {
> >>           "ceph version 14.2.18
> >> (befbc92f3c11eedd8626487211d200c0b44786d9) nautilus (stable)": 3
> >>       },
> >>       "osd": {
> >>           "ceph version 14.2.15
> >> (afdd217ae5fb1ed3f60e16bd62357ca58cc650e5) nautilus (stable)": 156
> >>       },
> >>       "mds": {
> >>           "ceph version 14.2.15
> >> (afdd217ae5fb1ed3f60e16bd62357ca58cc650e5) nautilus (stable)": 2
> >>       },
> >>       "overall": {
> >>           "ceph version 14.2.15
> >> (afdd217ae5fb1ed3f60e16bd62357ca58cc650e5) nautilus (stable)": 158,
> >>           "ceph version 14.2.18
> >> (befbc92f3c11eedd8626487211d200c0b44786d9) nautilus (stable)": 6
> >>       }
> >> }
> >>
> >>
> >> 12 OSDs are down
> >>
> >> # ceph -s
> >>     cluster:
> >>       id:     b489547c-ba50-4745-a914-23eb78e0e5dc
> >>       health: HEALTH_WARN
> >>               Reduced data availability: 5568 pgs inactive
> >>
> >>     services:
> >>       mon: 3 daemons, quorum cephmon3,cephmon1,cephmon2 (age 50m)
> >>       mgr: cephmon1(active, since 53m), standbys: cephmon3, cephmon2
> >>       mds: cephfs:1 {0=cephmds2=up:active} 1 up:standby
> >>       osd: 168 osds: 156 up (since 28m), 156 in (since 18m); 1722
> >> remapped pgs
> >>
> >>     data:
> >>       pools:   12 pools, 5568 pgs
> >>       objects: 0 objects, 0 B
> >>       usage:   0 B used, 0 B / 0 B avail
> >>       pgs:     100.000% pgs unknown
> >>                5568 unknown
> >>
> >>     progress:
> >>       Rebalancing after osd.103 marked in
> >>         [..............................]
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@xxxxxxx
> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux