Re: Troubleshooting cephadm - not deploying any daemons

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Have you checked /var/log/ceph/cephadm.log on the target nodes?

Zitat von "Zach Heise (SSCC)" <heise@xxxxxxxxxxxx>:

 Yes, sorry - I tried both 'ceph orch apply mgr "ceph01,ceph03"' and 'ceph orch apply mds "ceph04,ceph05"' before writing this initial email - once again, the same logged message: "6/8/22 2:25:12 PM[INF]Saving service mgr spec with placement ceph03;ceph01" but there's no messages logged about attempting to create the mgr daemon.

I tried this at the same time that I tried ''ceph orch apply mgr --placement=2' that I mentioned in my original email.

I think what I need is some advice on how to check cephadm's status - I assume it should be logging every time it tries to deploy a new daemon right? That should be my next stop, I think - looking at that log to see if it's even trying. I just don't know how to get to that point.

And it's not just mgr daemons, it's any kind of daemon so far, is not getting deployed.

But thank you for the advice, Dhairya.
-Zach

On 2022-06-08 3:44 PM, Dhairya Parmar wrote:
Hi Zach,

Try running `ceph orch apply mgr 2` or `ceph orch apply mgr --placement="<host1> <host2>"`. Refer this <https://docs.ceph.com/en/latest/cephadm/services/#orchestrator-cli-placement-spec> doc for more information, hope it helps.

Regards,
Dhairya

On Thu, Jun 9, 2022 at 1:59 AM Zach Heise (SSCC) <heise@xxxxxxxxxxxx> wrote:

   Our 16.2.7 cluster was deployed using cephadm from the start, but
   now it
   seems like deploying daemons with it is broken. Running 'ceph orch
   apply
   mgr --placement=2' causes '6/8/22 2:34:18 PM[INF]Saving service
   mgr spec
   with placement count:2' to appear in the logs, but a 2nd mgr does not
   get created.

   I also confirmed the same with mds daemons - using the dashboard, I
   tried creating a new set of MDS daemons "220606" count:3, but they
   never
   got deployed. The service type appears in the dashboard, though, just
   with no daemons deployed under it. Then I tried to delete it with the
   dashboard, and now 'ceph orch ls' outputs:

   NAME                       PORTS        RUNNING  REFRESHED AGE
   PLACEMENT
   mds.220606                                  0/3 <deleting> 15h 
   count:3

   More detail in YAML format doesn't even give me that much information:

   ceph01> ceph orch ls --service_name=mds.220606 --format yaml
   service_type: mds
   service_id: '220606'
   service_name: mds.220606
   placement:
      count: 3
   status:
      created: '2022-06-07T03:42:57.234124Z'
      running: 0
      size: 3
   events:
   - 2022-06-07T03:42:57.301349Z service:mds.220606 [INFO] "service was
   created"

   'ceph health detail' reports HEALTH_OK but cephadm doesn't seem to be
   doing its job. I read through the Cephadm troubleshooting page on
   ceph's
   website but since the daemons I'm trying to create don't even seem to
   try to spawn containers (podman ps shows the existing containers just
   fine) I don't know where to look next for logs, to see if cephadm +
   podman are trying to create new containers and failing, or not
   even trying.

   _______________________________________________
   ceph-users mailing list -- ceph-users@xxxxxxx
   To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux