Re: libceph: osdXXX up/down all the time

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Frank,

I asked the same question 4 years ago [1]. Basically, Gregs reponse was:

So, this is actually just noisy logging from the client processing an OSDMap. That should probably be turned down, as it's not really an indicator of...anything...as far as I can tell.

IIRC clients sometimes notice changes in the osdmap with some delay (if they didn't require to update the osdmap), so I just ignore these messages if the cluster is healthy otherwise and the clients work as expected. And here was my conclusion [2].

[1] https://www.spinics.net/lists/ceph-users/msg47279.html
[2] https://www.spinics.net/lists/ceph-users/msg47502.html

Zitat von Frank Schilder <frans@xxxxxx>:

Hi all,

on ceph fs kernel clients we see a lot of these kind of messages in bursts:

...
[Mon Dec 19 09:43:15 2022] libceph: osd1258 weight 0x10000 (in)
[Mon Dec 19 09:43:15 2022] libceph: osd1258 up
[Mon Dec 19 09:43:15 2022] libceph: osd1259 weight 0x10000 (in)
[Mon Dec 19 09:43:15 2022] libceph: osd1259 up
[Mon Dec 19 09:43:16 2022] libceph: osd0 down
[Mon Dec 19 09:43:16 2022] libceph: osd0 up
[Mon Dec 19 09:43:16 2022] libceph: osd0 down
[Mon Dec 19 09:43:16 2022] libceph: osd0 up
...

There haven't been osds going up and down since Wednesday last week. However, no OSDs went up/down. What is libceph actually reporting here?

The cluster is rebalancing since last Wednesday, when we added new disks. There has not been any daemon down since then:

# ceph status
  cluster:
    id:     ###
    health: HEALTH_OK

  services:
    mon: 5 daemons, quorum ceph-01,ceph-02,ceph-03,ceph-25,ceph-26 (age 12d)
mgr: ceph-25(active, since 11w), standbys: ceph-03, ceph-02, ceph-01, ceph-26
    mds: con-fs2:8 4 up:standby 8 up:active
    osd: 1260 osds: 1260 up (since 6d), 1260 in (since 6d); 2342 remapped pgs

  task status:

  data:
    pools:   14 pools, 25065 pgs
    objects: 1.53G objects, 2.8 PiB
    usage:   3.4 PiB used, 9.7 PiB / 13 PiB avail
    pgs:     1158282360/13309135349 objects misplaced (8.703%)
             22704 active+clean
             2261  active+remapped+backfill_wait
             81    active+remapped+backfilling
             16    active+clean+snaptrim
             3     active+clean+scrubbing+deep

  io:
    client:   120 MiB/s rd, 175 MiB/s wr, 1.46k op/s rd, 2.23k op/s wr
    recovery: 2.3 GiB/s, 850 objects/s

We are investigating the reason why some jobs on our HPC cluster get stuck after the job ends. These messages look somewhat suspicious and we wonder if it has anything to do with the ceph client/fs.

The cluster has been healthy the whole time.

Best regards and thanks for pointers!
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux