> -----Original Message----- > From: Jason Dillaman <jdillama@xxxxxxxxxx> > Sent: Friday, August 24, 2018 12:09 AM > To: sat <sat@xxxxxxxxxxxx> > Cc: ceph-users <ceph-users@xxxxxxxxxxxxxx> > Subject: Re: [question] one-way RBD mirroring doesn't work > > On Thu, Aug 23, 2018 at 10:56 AM sat <sat@xxxxxxxxxxxx> wrote: > > > > Hi, > > > > > > I'm trying to make a one-way RBD mirroed cluster between two Ceph > > clusters. But it hasn't worked yet. It seems to sucecss, but after > > making an RBD image from local cluster, it's considered as "unknown". > > > > ``` > > $ sudo rbd --cluster local create rbd/local.img --size=1G > > --image-feature=exclusive-lock,journaling > > $ sudo rbd --cluster local ls rbd > > local.img > > $ sudo rbd --cluster remote ls rbd > > local.img > > $ sudo rbd --cluster local mirror pool status rbd > > health: WARNING > > images: 1 total > > 1 unknown > > $ sudo rbd --cluster remote mirror pool status rbd > > health: OK > > images: 1 total > > 1 replaying > > $ > > ``` > > > > Could you tell me what is wrong? > > Nothing -- with one-directional RBD mirroring, on the receive side would report > status. If you started an rbd-mirror daemon against the "local" cluster, it would > report as healthy w/ that particular image in the "stopped" state since it's > primary. Thank you very much! So, since my images's state is "unknown" rather than "stopped", my rbd-mirror's setting is not correct. I'll chenk it. Best, Satoru > > > > > # detail > > > > There are two clusters, named "local" and "remote". "remote" is the mirror of > "local". > > Both two clusters has a pool, named "rbd". > > > > ## system environment > > > > - OS: ubuntu 16.04 > > - kernel: 4.4.0-112-generic > > - ceph: luminous 12.2.5 > > > > ## system configuration diagram > > > > > ================================================ > ============== > > +- manager(192.168.33.2): manipulate two clusters, > > | > > +- node0(192.168.33.3): "local"'s MON, MGR, and OSD0 > > | > > +- node1(192.168.33.4); "local"'s OSD1 > > | > > +- node2(192.168.33.5); "local"'s OSD2 > > | > > +- remote-node0(192.168.33.7): "remote"'s MON, MGR, OSD0, and > > +ceph-rbd-mirror > > | > > +- remote-node1(192.168.33.8); "remote"'s OSD1 > > | > > +- remote-node2(192.168.33.9); "remote"'s OSD2 > > > ================================================ > ================ > > > > # Step to reproduce > > > > 1. Prepare two clusters "local" and "remote" > > > > ``` > > $ sudo ceph --cluster local -s > > cluster: > > id: 9faca802-745d-43d8-b572-16617e553a5f > > health: HEALTH_WARN > > application not enabled on 1 pool(s) > > > > services: > > mon: 1 daemons, quorum 0 > > mgr: 0(active) > > osd: 3 osds: 3 up, 3 in > > > > data: > > pools: 1 pools, 128 pgs > > objects: 16 objects, 12395 kB > > usage: 3111 MB used, 27596 MB / 30708 MB avail > > pgs: 128 active+clean > > > > io: > > client: 852 B/s rd, 0 op/s rd, 0 op/s wr > > > > $ sudo ceph --cluster remote -s > > cluster: > > id: 1ecb0aa6-5a00-4946-bdba-bad78bfa4372 > > health: HEALTH_WARN > > application not enabled on 1 pool(s) > > > > services: > > mon: 1 daemons, quorum 0 > > mgr: 0(active) > > osd: 3 osds: 3 up, 3 in > > rbd-mirror: 1 daemon active > > > > data: > > pools: 1 pools, 128 pgs > > objects: 18 objects, 7239 kB > > usage: 3100 MB used, 27607 MB / 30708 MB avail > > pgs: 128 active+clean > > > > io: > > client: 39403 B/s rd, 0 B/s wr, 4 op/s rd, 0 op/s wr > > > > $ > > ```` > > > > > > Two clusters looks fine. > > > > 2. Setup one-way RBD pool mirroring from "local" to "remote" > > > > Setup an RBD pool mirroring between "local" and "remote" with the following > steps. > > > > > https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/3/h > > tml/block_device_guide/block_device_mirroring > > > > Both cluster's status look fine as follows. > > > > ``` > > $ sudo rbd --cluster local mirror pool info rbd > > Mode: pool > > Peers: none > > $ sudo rbd --cluster local mirror pool status rbd > > health: OK > > images: 0 total > > $ sudo rbd --cluster remote mirror pool info rbd > > Mode: pool > > Peers: > > UUID NAME CLIENT > > 53fb3a9a-c451-4552-b409-c08709ebe1a9 local client.local $ sudo rbd > > --cluster remote mirror pool status rbd > > health: OK > > images: 0 total > > $ > > ``` > > 3. Create an RBD image > > > > ``` > > $ sudo rbd --cluster local create rbd/local.img --size=1G > > --image-feature=exclusive-lock,journaling > > $ sudo rbd --cluster local ls rbd > > local.img > > $ sudo rbd --cluster remote ls rbd > > local.img > > $ > > ``` > > > > "rbd/local.img" seemd to be created and be mirrored fine. > > > > 4. Check both cluster's status and info > > > > Execute "rbd mirror pool info/status <pool>" and "info/status > > <rbd/local.img>" for both clusters. > > > > ## expected result > > > > Both "local" and "remote" report fine state. > > > > ## actual result > > > > Although "remote" works fine, but "local" reports seems not fine. > > > > ``` > > $ sudo rbd --cluster local mirror pool info rbd > > Mode: pool > > Peers: none > > $ sudo rbd --cluster local mirror pool status rbd > > health: WARNING > > images: 1 total > > 1 unknown > > $ sudo rbd --cluster local info rbd/local.img rbd image 'local.img': > > size 1024 MB in 256 objects > > order 22 (4096 kB objects) > > block_name_prefix: rbd_data.10336b8b4567 > > format: 2 > > features: exclusive-lock, journaling > > flags: > > create_timestamp: Mon Aug 20 06:01:29 2018 > > journal: 10336b8b4567 > > mirroring state: enabled > > mirroring global id: 447731a7-73ce-448d-90ac-38d05065f603 > > mirroring primary: true > > $ sudo rbd --cluster local status rbd/local.img > > Watchers: none > > $ sudo rbd --cluster remote mirror pool info rbd > > Mode: pool > > Peers: > > UUID NAME CLIENT > > 53fb3a9a-c451-4552-b409-c08709ebe1a9 local client.local $ sudo rbd > > --cluster remote mirror pool status rbd > > health: OK > > images: 1 total > > 1 replaying > > $ sudo rbd --cluster remote info rbd/local.img rbd image 'local.img': > > size 1024 MB in 256 objects > > order 22 (4096 kB objects) > > block_name_prefix: rbd_data.1025643c9869 > > format: 2 > > features: exclusive-lock, journaling > > flags: > > create_timestamp: Mon Aug 20 06:01:29 2018 > > journal: 1025643c9869 > > mirroring state: enabled > > mirroring global id: 447731a7-73ce-448d-90ac-38d05065f603 > > mirroring primary: false > > $ sudo rbd --cluster remote status rbd/local.img > > Watchers: > > watcher=192.168.33.7:0/506894125 client.4133 > > cookie=139703408874736 ``` > > > > Thanks, > > Satoru Takeuchi > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > -- > Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com