Hi, I'm trying to make a one-way RBD mirroed cluster between two Ceph clusters. But it hasn't worked yet. It seems to sucecss, but after making an RBD image from local cluster, it's considered as "unknown". ``` $ sudo rbd --cluster local create rbd/local.img --size=1G --image-feature=exclusive-lock,journaling $ sudo rbd --cluster local ls rbd local.img $ sudo rbd --cluster remote ls rbd local.img $ sudo rbd --cluster local mirror pool status rbd health: WARNING images: 1 total 1 unknown $ sudo rbd --cluster remote mirror pool status rbd health: OK images: 1 total 1 replaying $ ``` Could you tell me what is wrong? # detail There are two clusters, named "local" and "remote". "remote" is the mirror of "local". Both two clusters has a pool, named "rbd". ## system environment - OS: ubuntu 16.04 - kernel: 4.4.0-112-generic - ceph: luminous 12.2.5 ## system configuration diagram ============================================================== +- manager(192.168.33.2): manipulate two clusters, | +- node0(192.168.33.3): "local"'s MON, MGR, and OSD0 | +- node1(192.168.33.4); "local"'s OSD1 | +- node2(192.168.33.5); "local"'s OSD2 | +- remote-node0(192.168.33.7): "remote"'s MON, MGR, OSD0, and ceph-rbd-mirror | +- remote-node1(192.168.33.8); "remote"'s OSD1 | +- remote-node2(192.168.33.9); "remote"'s OSD2 ================================================================ # Step to reproduce 1. Prepare two clusters "local" and "remote" ``` $ sudo ceph --cluster local -s cluster: id: 9faca802-745d-43d8-b572-16617e553a5f health: HEALTH_WARN application not enabled on 1 pool(s) services: mon: 1 daemons, quorum 0 mgr: 0(active) osd: 3 osds: 3 up, 3 in data: pools: 1 pools, 128 pgs objects: 16 objects, 12395 kB usage: 3111 MB used, 27596 MB / 30708 MB avail pgs: 128 active+clean io: client: 852 B/s rd, 0 op/s rd, 0 op/s wr $ sudo ceph --cluster remote -s cluster: id: 1ecb0aa6-5a00-4946-bdba-bad78bfa4372 health: HEALTH_WARN application not enabled on 1 pool(s) services: mon: 1 daemons, quorum 0 mgr: 0(active) osd: 3 osds: 3 up, 3 in rbd-mirror: 1 daemon active data: pools: 1 pools, 128 pgs objects: 18 objects, 7239 kB usage: 3100 MB used, 27607 MB / 30708 MB avail pgs: 128 active+clean io: client: 39403 B/s rd, 0 B/s wr, 4 op/s rd, 0 op/s wr $ ```` Two clusters looks fine. 2. Setup one-way RBD pool mirroring from "local" to "remote" Setup an RBD pool mirroring between "local" and "remote" with the following steps. https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/3/html/block_device_guide/block_device_mirroring Both cluster's status look fine as follows. ``` $ sudo rbd --cluster local mirror pool info rbd Mode: pool Peers: none $ sudo rbd --cluster local mirror pool status rbd health: OK images: 0 total $ sudo rbd --cluster remote mirror pool info rbd Mode: pool Peers: UUID NAME CLIENT 53fb3a9a-c451-4552-b409-c08709ebe1a9 local client.local $ sudo rbd --cluster remote mirror pool status rbd health: OK images: 0 total $ ``` 3. Create an RBD image ``` $ sudo rbd --cluster local create rbd/local.img --size=1G --image-feature=exclusive-lock,journaling $ sudo rbd --cluster local ls rbd local.img $ sudo rbd --cluster remote ls rbd local.img $ ``` "rbd/local.img" seemd to be created and be mirrored fine. 4. Check both cluster's status and info Execute "rbd mirror pool info/status <pool>" and "info/status <rbd/local.img>" for both clusters. ## expected result Both "local" and "remote" report fine state. ## actual result Although "remote" works fine, but "local" reports seems not fine. ``` $ sudo rbd --cluster local mirror pool info rbd Mode: pool Peers: none $ sudo rbd --cluster local mirror pool status rbd health: WARNING images: 1 total 1 unknown $ sudo rbd --cluster local info rbd/local.img rbd image 'local.img': size 1024 MB in 256 objects order 22 (4096 kB objects) block_name_prefix: rbd_data.10336b8b4567 format: 2 features: exclusive-lock, journaling flags: create_timestamp: Mon Aug 20 06:01:29 2018 journal: 10336b8b4567 mirroring state: enabled mirroring global id: 447731a7-73ce-448d-90ac-38d05065f603 mirroring primary: true $ sudo rbd --cluster local status rbd/local.img Watchers: none $ sudo rbd --cluster remote mirror pool info rbd Mode: pool Peers: UUID NAME CLIENT 53fb3a9a-c451-4552-b409-c08709ebe1a9 local client.local $ sudo rbd --cluster remote mirror pool status rbd health: OK images: 1 total 1 replaying $ sudo rbd --cluster remote info rbd/local.img rbd image 'local.img': size 1024 MB in 256 objects order 22 (4096 kB objects) block_name_prefix: rbd_data.1025643c9869 format: 2 features: exclusive-lock, journaling flags: create_timestamp: Mon Aug 20 06:01:29 2018 journal: 1025643c9869 mirroring state: enabled mirroring global id: 447731a7-73ce-448d-90ac-38d05065f603 mirroring primary: false $ sudo rbd --cluster remote status rbd/local.img Watchers: watcher=192.168.33.7:0/506894125 client.4133 cookie=139703408874736 ``` Thanks, Satoru Takeuchi _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com