Re: RBD Mirroring down+unknown

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jan 7, 2020 at 11:17 AM <miguel.castillo@xxxxxxxxxx> wrote:
>
> Thanks for the reply Jason!
>
> We don't have selinux running on these machines, but I did fix the ownership on that file now so the ceph user can access it properly. The rbd mirror daemon does start up now, but the test image still shows down+unknown. I'll continue poking at it, but if you or anyone else can offer more things to look at and verify, that would be greatly appreciated!
>
> --------------------------------
> systemctl status ceph-rbd-mirror@admin
> ● ceph-rbd-mirror@admin.service - Ceph rbd mirror daemon
>    Loaded: loaded (/lib/systemd/system/ceph-rbd-mirror@.service; enabled; vendor preset: enabled)
>    Active: active (running) since Tue 2020-01-07 11:03:57 EST; 6min ago
>  Main PID: 917157 (rbd-mirror)
>     Tasks: 58
>    CGroup: /system.slice/system-ceph\x2drbd\x2dmirror.slice/ceph-rbd-mirror@admin.service
>            └─917157 /usr/bin/rbd-mirror -f --cluster ceph --id admin --setuser ceph --setgroup ceph
>
> Jan 07 11:03:57 ceph1-dc2 systemd[1]: Started Ceph rbd mirror daemon.
>
> # rbd --cluster dc1ceph mirror pool status fs_data --verbose

If the rbd-mirror daemon is running on ceph1-d2, you will need to run
the "rbd mirror pool status" command against that cluster. The future
Octopus release includes some improvements that replicate the
mirroring status between clusters so you would only need to check it
in one location (and designating RX/TX, RX-only, and TX-only
relationships), but in the meantime the status represents the
mirroring "pull" operation which in your case is occurring on DC2.

> health: WARNING
> images: 1 total
>     1 unknown
>
> mirror_test:
>   global_id:   c335017c-9b8f-49ee-9bc1-888789537c47
>   state:       down+unknown
>   description: status not found
>   last_update:
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx



-- 
Jason
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux