Re: Error ENOENT: Module not found

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Try the following command to fix the osd_remove_queue:

ceph config-key set mgr/cephadm/osd_remove_queue []

After that, the orchestrator will be back.

On Sat, Jan 25, 2025, 8:47 PM Devender Singh <devender@xxxxxxxxxx> wrote:

> +Eugen
> Lets follow “No recovery after removing node -
> active+undersized+degraded-- removed osd using purge…”.  Here.
>
> Sorry I missed ceph version which is 18.2.4. (with 5 nodes, 22osd each,
> where I removed one node and all mess.)
>
> Regards
> Dev
>
>
>
> > On Jan 25, 2025, at 11:34 AM, Devender Singh <devender@xxxxxxxxxx>
> wrote:
> >
> > Hello Fredreic
> >
> >
> > Thanks for your reply, Yes I also faced this issue after draining and
> removing of the node.
> > So used the same command and remove “original_weight” using ceph
> config-key get mgr/cephadm/osd_remove_queue and injected file again. Which
> resolved the orch issue.
> >
> > “Error ENOENT: Module not found - ceph orch commands stoppd working
> >
> > ceph config-key get mgr/cephadm/osd_remove_queue > osd_remove_queue.json
> >
> > Then only remove the "original_weight" key from that json and upload it
> back to the config-key store:
> >
> > ceph config-key set mgr/cephadm/osd_remove_queue -i
> osd_remove_queue_modified.json
> >
> > Then fail the mgr:
> >
> > ceph mgr fail”
> >
> >
> > But now issue is, my cluster showing objects misplaced, whereas I had 5
> nodes with host failure domain with R3 pool (size 3 and min 2), EC with
> 3+2.
> >
> > # ceph -s
> >   cluster:
> >     id:     384d7590-d018-11ee-b74c-5b2acfe0b35c
> >     health: HEALTH_WARN
> >             Degraded data redundancy: 2848547/29106793 objects degraded
> (9.787%), 105 pgs degraded, 132 pgs undersized
> >
> >   services:
> >     mon: 4 daemons, quorum node1,node5,node4,node2 (age 12h)
> >     mgr: node1.cvknae(active, since 12h), standbys: node4.foomun
> >     mds: 2/2 daemons up, 2 standby
> >     osd: 95 osds: 95 up (since 16h), 95 in (since 21h); 124 remapped pgs
> >     rgw: 2 daemons active (2 hosts, 1 zones)
> >
> >   data:
> >     volumes: 2/2 healthy
> >     pools:   18 pools, 817 pgs
> >     objects: 6.06M objects, 20 TiB
> >     usage:   30 TiB used, 302 TiB / 332 TiB avail
> >     pgs:     2848547/29106793 objects degraded (9.787%)
> >              2617833/29106793 objects misplaced (8.994%)
> >              561 active+clean
> >              124 active+clean+remapped
> >              105 active+undersized+degraded
> >              27  active+undersized
> >
> >   io:
> >     client:   1.4 MiB/s rd, 4.0 MiB/s wr, 25 op/s rd, 545 op/s wr
> >
> > And when using 'ceph config-key ls’ it’s showing old node and osd’s.
> >
> > # ceph config-key ls|grep -i 03n
> >     "config-history/135/+osd/host:node3/osd_memory_target",
> >     "config-history/14990/+osd/host:node3/osd_memory_target",
> >     "config-history/14990/-osd/host:node3/osd_memory_target",
> >     "config-history/15003/+osd/host:node3/osd_memory_target",
> >     "config-history/15003/-osd/host:node3/osd_memory_target",
> >     "config-history/15016/+osd/host:node3/osd_memory_target",
> >     "config-history/15016/-osd/host:node3/osd_memory_target",
> >     "config-history/15017/+osd/host:node3/osd_memory_target",
> >     "config-history/15017/-osd/host:node3/osd_memory_target",
> >     "config-history/15022/+osd/host:node3/osd_memory_target",
> >     "config-history/15022/-osd/host:node3/osd_memory_target",
> >     "config-history/15024/+osd/host:node3/osd_memory_target",
> >     "config-history/15024/-osd/host:node3/osd_memory_target",
> >     "config-history/15025/+osd/host:node3/osd_memory_target",
> >     "config-history/15025/-osd/host:node3/osd_memory_target",
> >     "config-history/153/+osd/host:node3/osd_memory_target",
> >     "config-history/153/-osd/host:node3/osd_memory_target",
> >     "config-history/165/+mon.node3/container_image",
> >     "config-history/171/-mon.node3/container_image",
> >     "config-history/176/+client.crash.node3/container_image",
> >     "config-history/182/-client.crash.node3/container_image",
> >     "config-history/4276/+osd/host:node3/osd_memory_target",
> >     "config-history/4276/-osd/host:node3/osd_memory_target",
> >     "config-history/433/+client.ceph-exporter.node3/container_image",
> >     "config-history/439/-client.ceph-exporter.node3/container_image",
> >     "config-history/459/+osd/host:node3/osd_memory_target",
> >     "config-history/459/-osd/host:node3/osd_memory_target",
> >     "config-history/465/+osd/host:node3/osd_memory_target",
> >     "config-history/465/-osd/host:node3/osd_memory_target",
> >     "config-history/4867/+osd/host:node3/osd_memory_target",
> >     "config-history/4867/-osd/host:node3/osd_memory_target",
> >     "config-history/4878/+mon.node3/container_image",
> >     "config-history/4884/-mon.node3/container_image",
> >     "config-history/4889/+client.crash.node3/container_image",
> >     "config-history/4895/-client.crash.node3/container_image",
> >
>  "config-history/5139/+mds.k8s-dev-cephfs.node3.iebxqn/container_image",
> >
>  "config-history/5142/-mds.k8s-dev-cephfs.node3.iebxqn/container_image",
> >     "config-history/5150/+client.ceph-exporter.node3/container_image",
> >     "config-history/5156/-client.ceph-exporter.node3/container_image",
> >     "config-history/5179/+osd/host:node3/osd_memory_target",
> >     "config-history/5179/-osd/host:node3/osd_memory_target",
> >     "config-history/5183/+client.rgw.sea-dev.node3.betyqd/rgw_frontends",
> >     "config-history/5189/+osd/host:node3/osd_memory_target",
> >     "config-history/5189/-osd/host:node3/osd_memory_target",
> >     "config-history/6929/-client.rgw.sea-dev.node3.betyqd/rgw_frontends",
> >     "config-history/6933/+osd/host:node3/osd_memory_target",
> >     "config-history/6933/-osd/host:node3/osd_memory_target",
> >     "config-history/9710/+osd/host:node3/osd_memory_target",
> >     "config-history/9710/-osd/host:node3/osd_memory_target",
> >     "config/osd/host:node3/osd_memory_target”,
> >
> >
> > Regards
> > Dev
> >
> >> On Jan 25, 2025, at 4:39 AM, Frédéric Nass <
> frederic.nass@xxxxxxxxxxxxxxxx> wrote:
> >>
> >> Hi,
> >>
> >> I've seen this happening on a test cluster after draining a host that
> also had a MGR service. Can you check if Eugen's solution here [1] helps in
> your case ? And maybe investigate 'ceph config-key ls' for any issues in
> config keys ?
> >>
> >> Regards,
> >> Frédéric.
> >>
> >> [1] https://www.spinics.net/lists/ceph-users/msg83667.html <
> https://www.google.com/url?q=https://www.spinics.net/lists/ceph-users/msg83667.html&source=gmail-imap&ust=1738413580000000&usg=AOvVaw3Zk70LrQ6SrLX02gJ7Cowl
> >
> >>
> >> De : Devender Singh <devender@xxxxxxxxxx>
> >> Envoyé : samedi 25 janvier 2025 06:27
> >> À : Fnu Virender Kumar
> >> Cc: ceph-users
> >> Objet :  Re: Error ENOENT: Module not found
> >>
> >> Thanks for you reply… but those command not working as its an always
> module..but strange still showing error,
> >>
> >> # ceph  mgr module enable orchestrator
> >> module 'orchestrator' is already enabled (always-on)
> >>
> >> # ceph orch set backend  — returns successfully…
> >>
> >> # # ceph orch ls
> >> Error ENOENT: No orchestrator configured (try `ceph orch set backend`)
> >>
> >> Its revolving between same error..
> >>
> >> Root Cause: I removed a hosts and its odd’s and after some time above
> error started automatically.
> >>
> >> Earlier in the had 5  nodes  but now 4.. Cluster is showing  unclean pg
> but not doing anything..
> >>
> >> But big error is Error ENOENT:
> >>
> >>
> >> Regards
> >> Dev
> >>
> >> > On Jan 24, 2025, at 4:59 PM, Fnu Virender Kumar <
> virenderk@xxxxxxxxxxxx> wrote:
> >> >
> >> > Did you try
> >> >
> >> > Ceph mgr module enable orchestrator
> >> > Ceph orch set backend
> >> > Ceph orch ls
> >> >
> >> > Check the mgr service daemon as well
> >> > Ceph -s
> >> >
> >> >
> >> > Regards
> >> > Virender
> >> > From: Devender Singh <devender@xxxxxxxxxx>
> >> > Sent: Friday, January 24, 2025 6:34:43 PM
> >> > To: ceph-users <ceph-users@xxxxxxx>
> >> > Subject:  Error ENOENT: Module not found
> >> >
> >> >
> >> > Hello all
> >> >
> >> > Any quick fix for …
> >> >
> >> > root@sea-devnode1:~# ceph orch ls
> >> > Error ENOENT: Module not found
> >> >
> >> >
> >> > Regards
> >> > Dev
> >> > _______________________________________________
> >> > ceph-users mailing list -- ceph-users@xxxxxxx
> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> >> _______________________________________________
> >> ceph-users mailing list -- ceph-users@xxxxxxx
> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux