Hi, That's probably related to https://tracker.ceph.com/issues/51571 Regards, Dimitri On Wed, Jul 14, 2021 at 8:17 AM Eugen Block <eblock@xxxxxx> wrote: > Hi, > > do you see the daemon on that iscsi host(s) with 'cephadm ls'? If the > answer is yes, you could remove it with cephadm, too: > > cephadm rm-daemon --name iscsi.iscsi > > Does that help? > > > Zitat von Fyodor Ustinov <ufm@xxxxxx>: > > > Hi! > > > > I have fresh installed pacific > > > > root@s-26-9-19-mon-m1:~# ceph version > > ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a) > > pacific (stable) > > > > I managed to bring him to this state: > > > > root@s-26-9-19-mon-m1:~# ceph health detail > > HEALTH_ERR Module 'cephadm' has failed: dashboard iscsi-gateway-rm > > failed: iSCSI gateway 'iscsi-gw-1' does not exist retval: -2 > > [ERR] MGR_MODULE_ERROR: Module 'cephadm' has failed: dashboard > > iscsi-gateway-rm failed: iSCSI gateway 'iscsi-gw-1' does not exist > > retval: -2 > > Module 'cephadm' has failed: dashboard iscsi-gateway-rm failed: > > iSCSI gateway 'iscsi-gw-1' does not exist retval: -2 > > > > > > root@s-26-9-19-mon-m1:~# ceph orch ls > > NAME PORTS RUNNING REFRESHED AGE PLACEMENT > > alertmanager ?:9093,9094 1/1 14m ago 9d count:1;label:mon > > crash 12/12 14m ago 11d * > > grafana ?:3000 1/1 14m ago 9d count:1;label:mon > > iscsi.iscsi 0/0 <deleting> 7h > iscsi-gw-1;iscsi-gw-2 > > mgr 2/2 14m ago 9d count:2;label:mon > > mon 3/3 14m ago 5d count:3 > > node-exporter ?:9100 12/12 14m ago 11d * > > osd 54/54 14m ago - <unmanaged> > > prometheus ?:9095 1/1 14m ago 5d count:1;label:mon > > > > root@s-26-9-19-mon-m1:~# ceph orch host ls > > HOST ADDR LABELS STATUS > > s-26-9-17 10.5.107.104 _admin > > s-26-9-18 10.5.107.105 _admin > > s-26-9-19-mon-m1 10.5.107.101 mon _admin > > s-26-9-20 10.5.107.106 _admin > > s-26-9-21 10.5.107.107 _admin > > s-26-9-22 10.5.107.110 _admin > > s-26-9-23 10.5.107.108 _admin > > s-26-9-24-mon-m2 10.5.107.102 _admin mon > > s-26-9-25 10.5.107.111 _admin > > s-26-9-26 10.5.107.109 _admin > > s-26-9-27 10.5.107.112 _admin > > s-26-9-28-mon-m3 10.5.107.103 _admin mon > > > > > > How can we get the cluster out of this state now? > > > > WBR, > > Fyodor. > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx