Hi, Note : Firewall is disabled on all hosts. Regards. Le ven. 15 mars 2024 à 06:42, wodel youchi <wodel.youchi@xxxxxxxxx> a écrit : > Hi, > > I did recreate the cluster again, and it is the result. > > This is my initial bootstrap > > cephadm --image 192.168.2.36:4000/ceph/ceph:v18 bootstrap > --initial-dashboard-user admin \ > --initial-dashboard-password adminpass --dashboard-password-noupdate > --registry-url 192.168.2.36:4000 \ > --registry-username admin --registry-password admin --mon-ip 20.1.0.23 > --cluster-network 20.2.0.0/16 \ > --ssh-private-key /root/.ssh/id_rsa --ssh-public-key /root/.ssh/id_rsa.pub > \ > -c initial-ceph.conf > > This is my initial-ceph.conf > [mgr] > mgr/cephadm/container_image_prometheus = > 192.168.2.36:4000/prometheus/prometheus:v2.43.0 > mgr/cephadm/container_image_node_exporter = > 192.168.2.36:4000/prometheus/node-exporter:v1.5.0 > mgr/cephadm/container_image_grafana = > 192.168.2.36:4000/ceph/ceph-grafana:9.4.7 > mgr/cephadm/container_image_alertmanager = > 192.168.2.36:4000/prometheus/alertmanager:v0.25.0 > > > Then I added two managers and monitors > # ceph orch host add controllerb 20.1.0.27 _admin > # ceph orch host add controllerc 20.1.0.31 _admin > # ceph orch apply mon --placement="3 controllera controllerb controllerc" > # ceph orch apply mgr --placement="3 controllera controllerb controllerc" > > Then I added node-exporter, prometheus, grafana and crash > Then I added osd hosts > # ceph orch host add computehci01 20.1.0.2 > # ceph orch host add computehci02 20.1.0.3 > # ceph orch host add computehci03 20.1.0.4 > ... > # ceph orch host add computehci09 20.1.0.10 > ... > > And finally I added osd daemons > # ceph orch daemon add osd > computehci01:/dev/nvme0n1,/dev/nvme1n1,/dev/nvme2n1,/dev/nvme3n1 > # ceph orch daemon add osd > computehci02:/dev/nvme0n1,/dev/nvme1n1,/dev/nvme2n1,/dev/nvme3n1 > ... > > I created a pool > # ceph osd pool create volumes replicated > # ceph osd pool application enable volumes rbd > > I even created cephfs pools and added mds service, but still 100% pgs are > unknown???? !!!! > > > [root@controllera ~]# ceph -s > cluster: > id: df914aa2-e21a-11ee-b8df-3cecef2872f0 > health: HEALTH_WARN > 1 MDSs report slow metadata IOs > Reduced data availability: 4 pgs inactive > > services: > mon: 3 daemons, quorum controllera,controllerc,controllerb (age 13h) > mgr: controllera.ajttxz(active, since 13h), standbys: > controllerb.qtixeq, controllerc.pqyqqo > mds: 1/1 daemons up, 2 standby > osd: 36 osds: 36 up (since 7h), 36 in (since 7h) > > data: > volumes: 1/1 healthy > pools: 4 pools, 4 pgs > objects: 0 objects, 0 B > usage: 1.1 GiB used, 110 TiB / 110 TiB avail > pgs: 100.000% pgs unknown > 4 unknown > > [root@controllera ~]# ceph osd pool ls detail > pool 1 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 46 flags > hashpspool,creating stripe_width 0 pg_num_max 32 pg_nu > m_min 1 application mgr > pool 2 'volumes' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 154 flags > hashpspool,creating stripe_width 0 application rbd > pool 3 'cephfs' replicated size 3 min_size 2 crush_rule 0 object_hash > rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 157 flags > hashpspool,creating stripe_width 0 application ceph > fs > pool 4 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0 > object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 157 > flags hashpspool,creating stripe_width 0 pg_auto > scale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs > > What am I missing, why PGs won't pair? > > > > Regards. > > Le jeu. 14 mars 2024 à 15:36, wodel youchi <wodel.youchi@xxxxxxxxx> a > écrit : > >> Hi, >> >> I am creating a new ceph cluster using REEF. >> >> This is my host_specs file >> [root@controllera config]# cat hosts-specs2.yml >> service_type: host >> hostname: computehci01 >> addr: 20.1.0.2 >> location: >> chassis: chassis1 >> --- >> service_type: host >> hostname: computehci02 >> addr: 20.1.0.3 >> location: >> chassis: chassis1 >> --- >> service_type: host >> hostname: computehci03 >> addr: 20.1.0.4 >> location: >> chassis: chassis1 >> --- >> service_type: host >> hostname: computehci04 >> addr: 20.1.0.5 >> location: >> chassis: chassis2 >> --- >> service_type: host >> hostname: computehci05 >> addr: 20.1.0.6 >> location: >> chassis: chassis2 >> --- >> service_type: host >> hostname: computehci06 >> addr: 20.1.0.7 >> location: >> chassis: chassis2 >> --- >> service_type: host >> hostname: computehci07 >> addr: 20.1.0.8 >> location: >> chassis: chassis3 >> --- >> service_type: host >> hostname: computehci08 >> addr: 20.1.0.9 >> location: >> chassis: chassis3 >> --- >> service_type: host >> hostname: computehci09 >> addr: 20.1.0.10 >> location: >> chassis: chassis3 >> --- >> service_type: host >> hostname: computehci10 >> addr: 20.1.0.11 >> location: >> chassis: chassis3 >> --- >> service_type: host >> hostname: computehci11 >> addr: 20.1.0.12 >> location: >> chassis: chassis4 >> --- >> service_type: host >> hostname: computehci12 >> addr: 20.1.0.13 >> location: >> chassis: chassis4 >> --- >> service_type: host >> hostname: computehci13 >> addr: 20.1.0.14 >> location: >> chassis: chassis4 >> --- >> service_type: host >> hostname: computehci14 >> addr: 20.1.0.15 >> location: >> chassis: chassis4 >> --- >> service_type: host >> hostname: computehci15 >> addr: 20.1.0.16 >> location: >> chassis: chassis5 >> --- >> service_type: host >> hostname: computehci16 >> addr: 20.1.0.17 >> location: >> chassis: chassis5 >> --- >> service_type: host >> hostname: computehci17 >> addr: 20.1.0.18 >> location: >> chassis: chassis5 >> --- >> service_type: host >> hostname: computehci18 >> addr: 20.1.0.19 >> location: >> chassis: chassis5 >> --- >> service_type: host >> hostname: computehci19 >> addr: 20.1.0.20 >> location: >> chassis: chassis6 >> --- >> service_type: host >> hostname: computehci20 >> addr: 20.1.0.21 >> location: >> chassis: chassis6 >> --- >> service_type: host >> hostname: computehci21 >> addr: 20.1.0.22 >> location: >> chassis: chassis6 >> --- >> service_type: host >> hostname: computehci22 >> addr: 20.1.0.24 >> location: >> chassis: chassis7 >> --- >> service_type: host >> hostname: computehci23 >> addr: 20.1.0.25 >> location: >> chassis: chassis7 >> --- >> service_type: host >> hostname: computehci24 >> addr: 20.1.0.26 >> location: >> chassis: chassis7 >> --- >> service_type: host >> hostname: computehci25 >> addr: 20.1.0.28 >> location: >> chassis: chassis8 >> --- >> service_type: host >> hostname: computehci26 >> addr: 20.1.0.29 >> location: >> chassis: chassis8 >> --- >> service_type: host >> hostname: computehci27 >> addr: 20.1.0.30 >> location: >> chassis: chassis8 >> --- >> service_type: host >> hostname: controllera >> addr: 20.1.0.23 >> --- >> service_type: host >> hostname: controllerb >> addr: 20.1.0.27 >> --- >> service_type: host >> hostname: controllerc >> addr: 20.1.0.31 >> --- >> service_type: mon >> placement: >> hosts: >> - controllera >> - controllerb >> - controllerc >> --- >> service_type: mgr >> placement: >> hosts: >> - controllera >> - controllerb >> - controllerc >> --- >> service_type: osd >> service_id: default_drive_group >> placement: >> hosts: >> - computehci01 >> - computehci02 >> - computehci03 >> - computehci04 >> - computehci05 >> - computehci06 >> - computehci07 >> - computehci08 >> - computehci09 >> - computehci10 >> - computehci11 >> - computehci12 >> - computehci13 >> - computehci14 >> - computehci15 >> - computehci16 >> - computehci17 >> - computehci18 >> - computehci19 >> - computehci20 >> - computehci21 >> - computehci22 >> - computehci23 >> - computehci24 >> - computehci25 >> - computehci26 >> - computehci27 >> spec: >> data_devices: >> rotational: 0 >> >> >> All osds were added but, pg still unknown state >> >> I've created a pool, but it didn't change anything. >> >> [root@controllerb ~]# ceph -s >> cluster: >> id: be250ade-e1f2-11ee-a6ff-3cecef2872f0 >> health: HEALTH_WARN >> Reduced data availability: 1 pg inactive >> >> services: >> mon: 3 daemons, quorum controllera,controllerc,controllerb (age 3h) >> mgr: controllerc.jevbkl(active, since 21s), standbys: >> controllera.zwlolp, controllerb.vqkdga >> osd: 108 osds: 108 up (since 2m), 108 in (since 24m) >> >> data: >> pools: 2 pools, 33 pgs >> objects: 0 objects, 0 B >> usage: 5.1 GiB used, 330 TiB / 330 TiB avail >> >> >> * pgs: 100.000% pgs unknown 33 unknown* >> >> Did I miss something? >> >> Regards. >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx