Re: [REEF][cephadm] new cluster all pg unknown

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I did recreate the cluster again, and it is the result.

This is my initial bootstrap

cephadm --image 192.168.2.36:4000/ceph/ceph:v18 bootstrap
--initial-dashboard-user admin \
--initial-dashboard-password adminpass --dashboard-password-noupdate
 --registry-url 192.168.2.36:4000  \
--registry-username admin --registry-password admin --mon-ip 20.1.0.23
--cluster-network 20.2.0.0/16 \
--ssh-private-key /root/.ssh/id_rsa --ssh-public-key /root/.ssh/id_rsa.pub \
-c initial-ceph.conf

This is my initial-ceph.conf
[mgr]
mgr/cephadm/container_image_prometheus =
192.168.2.36:4000/prometheus/prometheus:v2.43.0
mgr/cephadm/container_image_node_exporter =
192.168.2.36:4000/prometheus/node-exporter:v1.5.0
mgr/cephadm/container_image_grafana =
192.168.2.36:4000/ceph/ceph-grafana:9.4.7
mgr/cephadm/container_image_alertmanager =
192.168.2.36:4000/prometheus/alertmanager:v0.25.0


Then I added two managers and monitors
# ceph orch host add controllerb 20.1.0.27 _admin
# ceph orch host add controllerc 20.1.0.31 _admin
# ceph orch apply mon --placement="3 controllera controllerb controllerc"
# ceph orch apply mgr --placement="3 controllera controllerb controllerc"

Then I added node-exporter, prometheus, grafana and crash
Then I added osd hosts
# ceph orch host add computehci01 20.1.0.2
# ceph orch host add computehci02 20.1.0.3
# ceph orch host add computehci03 20.1.0.4
...
# ceph orch host add computehci09 20.1.0.10
...

And finally I added osd daemons
# ceph orch daemon add osd
computehci01:/dev/nvme0n1,/dev/nvme1n1,/dev/nvme2n1,/dev/nvme3n1
# ceph orch daemon add osd
computehci02:/dev/nvme0n1,/dev/nvme1n1,/dev/nvme2n1,/dev/nvme3n1
...

I created a pool
# ceph osd pool create volumes  replicated
# ceph osd pool application enable volumes rbd

I even created cephfs pools and added mds service, but still 100% pgs are
unknown???? !!!!


[root@controllera ~]# ceph -s
 cluster:
   id:     df914aa2-e21a-11ee-b8df-3cecef2872f0
   health: HEALTH_WARN
           1 MDSs report slow metadata IOs
           Reduced data availability: 4 pgs inactive

 services:
   mon: 3 daemons, quorum controllera,controllerc,controllerb (age 13h)
   mgr: controllera.ajttxz(active, since 13h), standbys:
controllerb.qtixeq, controllerc.pqyqqo
   mds: 1/1 daemons up, 2 standby
   osd: 36 osds: 36 up (since 7h), 36 in (since 7h)

 data:
   volumes: 1/1 healthy
   pools:   4 pools, 4 pgs
   objects: 0 objects, 0 B
   usage:   1.1 GiB used, 110 TiB / 110 TiB avail
   pgs:     100.000% pgs unknown
            4 unknown

[root@controllera ~]# ceph osd pool ls detail
pool 1 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 46 flags
hashpspool,creating stripe_width 0 pg_num_max 32 pg_nu
m_min 1 application mgr
pool 2 'volumes' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 154 flags
hashpspool,creating stripe_width 0 application rbd
pool 3 'cephfs' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 157 flags
hashpspool,creating stripe_width 0 application ceph
fs
pool 4 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 157
flags hashpspool,creating stripe_width 0 pg_auto
scale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs

What am I missing, why PGs won't pair?



Regards.

Le jeu. 14 mars 2024 à 15:36, wodel youchi <wodel.youchi@xxxxxxxxx> a
écrit :

> Hi,
>
> I am creating a new ceph cluster using REEF.
>
> This is my host_specs file
> [root@controllera config]# cat hosts-specs2.yml
> service_type: host
> hostname: computehci01
> addr: 20.1.0.2
> location:
>  chassis: chassis1
> ---
> service_type: host
> hostname: computehci02
> addr: 20.1.0.3
> location:
>  chassis: chassis1
> ---
> service_type: host
> hostname: computehci03
> addr: 20.1.0.4
> location:
>  chassis: chassis1
> ---
> service_type: host
> hostname: computehci04
> addr: 20.1.0.5
> location:
>  chassis: chassis2
> ---
> service_type: host
> hostname: computehci05
> addr: 20.1.0.6
> location:
>  chassis: chassis2
> ---
> service_type: host
> hostname: computehci06
> addr: 20.1.0.7
> location:
>  chassis: chassis2
> ---
> service_type: host
> hostname: computehci07
> addr: 20.1.0.8
> location:
>  chassis: chassis3
> ---
> service_type: host
> hostname: computehci08
> addr: 20.1.0.9
> location:
>  chassis: chassis3
> ---
> service_type: host
> hostname: computehci09
> addr: 20.1.0.10
> location:
>  chassis: chassis3
> ---
> service_type: host
> hostname: computehci10
> addr: 20.1.0.11
> location:
>  chassis: chassis3
> ---
> service_type: host
> hostname: computehci11
> addr: 20.1.0.12
> location:
>  chassis: chassis4
> ---
> service_type: host
> hostname: computehci12
> addr: 20.1.0.13
> location:
>  chassis: chassis4
> ---
> service_type: host
> hostname: computehci13
> addr: 20.1.0.14
> location:
>  chassis: chassis4
> ---
> service_type: host
> hostname: computehci14
> addr: 20.1.0.15
> location:
>  chassis: chassis4
> ---
> service_type: host
> hostname: computehci15
> addr: 20.1.0.16
> location:
>  chassis: chassis5
> ---
> service_type: host
> hostname: computehci16
> addr: 20.1.0.17
> location:
>  chassis: chassis5
> ---
> service_type: host
> hostname: computehci17
> addr: 20.1.0.18
> location:
>  chassis: chassis5
> ---
> service_type: host
> hostname: computehci18
> addr: 20.1.0.19
> location:
>  chassis: chassis5
> ---
> service_type: host
> hostname: computehci19
> addr: 20.1.0.20
> location:
>  chassis: chassis6
> ---
> service_type: host
> hostname: computehci20
> addr: 20.1.0.21
> location:
>  chassis: chassis6
> ---
> service_type: host
> hostname: computehci21
> addr: 20.1.0.22
> location:
>  chassis: chassis6
> ---
> service_type: host
> hostname: computehci22
> addr: 20.1.0.24
> location:
>  chassis: chassis7
> ---
> service_type: host
> hostname: computehci23
> addr: 20.1.0.25
> location:
>  chassis: chassis7
> ---
> service_type: host
> hostname: computehci24
> addr: 20.1.0.26
> location:
>  chassis: chassis7
> ---
> service_type: host
> hostname: computehci25
> addr: 20.1.0.28
> location:
>  chassis: chassis8
> ---
> service_type: host
> hostname: computehci26
> addr: 20.1.0.29
> location:
>  chassis: chassis8
> ---
> service_type: host
> hostname: computehci27
> addr: 20.1.0.30
> location:
>  chassis: chassis8
> ---
> service_type: host
> hostname: controllera
> addr: 20.1.0.23
> ---
> service_type: host
> hostname: controllerb
> addr: 20.1.0.27
> ---
> service_type: host
> hostname: controllerc
> addr: 20.1.0.31
> ---
> service_type: mon
> placement:
>  hosts:
>   - controllera
>   - controllerb
>   - controllerc
> ---
> service_type: mgr
> placement:
>  hosts:
>   - controllera
>   - controllerb
>   - controllerc
> ---
> service_type: osd
> service_id: default_drive_group
> placement:
>  hosts:
>   - computehci01
>   - computehci02
>   - computehci03
>   - computehci04
>   - computehci05
>   - computehci06
>   - computehci07
>   - computehci08
>   - computehci09
>   - computehci10
>   - computehci11
>   - computehci12
>   - computehci13
>   - computehci14
>   - computehci15
>   - computehci16
>   - computehci17
>   - computehci18
>   - computehci19
>   - computehci20
>   - computehci21
>   - computehci22
>   - computehci23
>   - computehci24
>   - computehci25
>   - computehci26
>   - computehci27
> spec:
>  data_devices:
>    rotational: 0
>
>
> All osds were added but, pg still unknown state
>
> I've created a pool, but it didn't change anything.
>
> [root@controllerb ~]# ceph -s
>  cluster:
>    id:     be250ade-e1f2-11ee-a6ff-3cecef2872f0
>    health: HEALTH_WARN
>            Reduced data availability: 1 pg inactive
>
>  services:
>    mon: 3 daemons, quorum controllera,controllerc,controllerb (age 3h)
>    mgr: controllerc.jevbkl(active, since 21s), standbys:
> controllera.zwlolp, controllerb.vqkdga
>    osd: 108 osds: 108 up (since 2m), 108 in (since 24m)
>
>  data:
>    pools:   2 pools, 33 pgs
>    objects: 0 objects, 0 B
>    usage:   5.1 GiB used, 330 TiB / 330 TiB avail
>
>
> *  pgs:     100.000% pgs unknown             33 unknown*
>
> Did I miss something?
>
> Regards.
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux