Hi Jason,
I'm using nbd-rbd.
Under /var/log/ceph/client.log
I see this error:
2017-06-25 05:25:32.833202 7f658ff04e00 0 ceph version 10.2.7
(50e863e0f4bc8f4b9e31156de690d765af245185), process rbd-nbd, pid 8524
2017-06-25 05:25:32.853811 7f658ff04e00 -1 asok(0x7f6599531900)
AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed
to bind the UNIX domain socket to
'/var/run/ceph/ceph-client.admin.asok': (17) File exists
Il 25/06/2017 15:00, Jason Dillaman ha scritto:
Are you using librbd via QEMU or krbd? If librbd, what errors are
noted in the instance's librbd log file?
On Sun, Jun 25, 2017 at 4:30 AM, Massimiliano Cuttini <max@xxxxxxxxxxxxx> wrote:
After 4 months of test we decided to go live and store real VDI in
production.
However just the same day something went suddenly wrong.
The last copy of the VDI in Ceph was corrupted.
Trying to fix the filesystem works for open it, but mysqld never went online
even after reinstalling but only forcing fixing of innoDB.
To many database corrupted. Finally the last reboot VDI became inaccessible
and do not start.
At the sametime other 3 VDI stuck and becomes inaccessible, try to access to
them get this error:
cannot open /dev/xvdb: Input/output error
If I run from the Dom(0) the ceph -s it seems allright:
cluster 33693979-3177-44f4-bcb8-1b3bbc658352
health HEALTH_OK
monmap e2: 2 mons at
{ceph-node1=xxxxxx:6789/0,ceph-node2=yyyyyy:6789/0}
election epoch 218, quorum 0,1 ceph-node1,ceph-node2
osdmap e2560: 8 osds: 8 up, 8 in
flags sortbitwise,require_jewel_osds
pgmap v8311677: 484 pgs, 3 pools, 940 GB data, 444 kobjects
1883 GB used, 5525 GB / 7408 GB avail
484 active+clean
client io 482 kB/s wr, 0 op/s rd, 5 op/s wr
Everything seems ok.
I can see pool and rbd:
NAME
SIZE PARENT FMT PROT LOCK
VHD-1649fde4-6637-43d8-b815-656e4080887d
102400M 2
VHD-1bcd9d72-d8fe-4fc2-ad6a-ce84b18e4340
102400M 2
VHD-239ea931-5ddd-4aaf-bc89-b192641f6dcf
200G 2
VHD-3e16395d-7dad-4680-a7ad-7f398da7fd9e
200G 2
VHD-41a76fe7-c9ff-4082-adb4-43f3120a9106
102400M 2
VHD-41a76fe7-c9ff-4082-adb4-43f3120a9106@SNAP-346d01f4-cbd4-4e8a-af8f-473c2de3c60d
102400M 2 yes
VHD-48fdb12d-110b-419c-9330-0f05827fe41e
102400M 2
VHD-602b05be-395d-442e-bd68-7742deaf97bd
200G 2
VHD-691e4fb4-5f7b-4bc1-af7b-98780d799067
102400M 2
VHD-6da2154e-06fd-4063-8af5-ae86ae61df50
10240M 2 excl
VHD-8631ab86-c85c-407b-9e15-bd86e830ba74
200G 2
VHD-97cbc3d2-519c-4d11-b795-7a71179e193c
102400M 2
VHD-97cbc3d2-519c-4d11-b795-7a71179e193c@SNAP-ad9ab028-57d9-48e4-9d13-a9a02b06e68a
102400M 2 yes
VHD-acb2a9b0-e98d-474e-aa42-ed4e5534ddbe
102400M 2 excl
VHD-adaa639e-a7a4-4723-9c18-0c2b3ab9d99f
102400M 2
VHD-bb0f40c7-206b-4023-b56f-4a70d02d5a58
200G 2
VHD-bb0f40c7-206b-4023-b56f-4a70d02d5a58@SNAP-f8b1cc40-4a9b-4550-8bfa-7fad467e726b
200G 2 yes
VHD-c8aca7bd-1e37-4af4-b642-f267602e210f
102400M 2
VHD-c8aca7bd-1e37-4af4-b642-f267602e210f@SNAP-0e830e68-c7b4-49a9-8210-4c0b4ecca762
102400M 2 yes
VHD-cf2139ac-b1c4-404d-87da-db8f992a3e72
102400M 2
VHD-cf2139ac-b1c4-404d-87da-db8f992a3e72@SNAP-4fb59ee5-3256-4e75-9d48-facdd818d754
102400M 2 yes
however they are unbootable and unreadable even as secondary driver.
Hypervisor refure to load them in the VMs.
What Can I do?
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com