i have a file which is untouchable: ls -i gives an error, stat gives an error. it shows ??? for all fields except name.
How do i clean this up?
I'm on ubuntu 15.10, running 0.94.5
# ceph -v
ceph version 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43)
the node that accessed the file then caused a problem with mds:
root@nubo-1:/home/git/go/src/github.com/gogits/gogs# ceph status
cluster b23abffc-71c4-4464-9449-3f2c9fbe1ded
health HEALTH_WARN
mds0: Client nubo-1 failing to respond to capability release
monmap e1: 3 mons at {nubo-1=10.100.10.60:6789/0,nubo-2=10.100.10.61:6789/0,nubo-3=10.100.10.62:6789/0}
election epoch 906, quorum 0,1,2 nubo-1,nubo-2,nubo-3
mdsmap e418: 1/1/1 up {0=nubo-2=up:active}, 2 up:standby
osdmap e2081: 6 osds: 6 up, 6 in
pgmap v95696: 560 pgs, 6 pools, 131 GB data, 97784 objects
265 GB used, 5357 GB / 5622 GB avail
560 active+clean
Trying a different node, i see the same problem.
I'm getting this error dumped to dmesg:
[670243.421212] Workqueue: ceph-msgr con_work [libceph]
[670243.421213] 0000000000000000 00000000e800e516 ffff8810cd68f9d8 ffffffff817e8c09
[670243.421215] 0000000000000000 0000000000000000 ffff8810cd68fa18 ffffffff8107b3c6
[670243.421217] ffff8810cd68fa28 00000000ffffffea 0000000000000000 0000000000000000
[670243.421218] Call Trace:
[670243.421221] [<ffffffff817e8c09>] dump_stack+0x45/0x57
[670243.421223] [<ffffffff8107b3c6>] warn_slowpath_common+0x86/0xc0
[670243.421225] [<ffffffff8107b4fa>] warn_slowpath_null+0x1a/0x20
[670243.421229] [<ffffffffc06ebb1c>] fill_inode.isra.18+0xc5c/0xc90 [ceph]
[670243.421233] [<ffffffff81217427>] ? inode_init_always+0x107/0x1b0
[670243.421236] [<ffffffffc06e95e0>] ? ceph_mount+0x7e0/0x7e0 [ceph]
[670243.421241] [<ffffffffc06ebe82>] ceph_fill_trace+0x332/0x910 [ceph]
[670243.421248] [<ffffffffc0709db5>] handle_reply+0x525/0xb70 [ceph]
[670243.421255] [<ffffffffc070cac8>] dispatch+0x3c8/0xbb0 [ceph]
[670243.421260] [<ffffffffc069daeb>] con_work+0x57b/0x1770 [libceph]
[670243.421262] [<ffffffff810b2d7b>] ? dequeue_task_fair+0x36b/0x700
[670243.421263] [<ffffffff810b2141>] ? put_prev_entity+0x31/0x420
[670243.421265] [<ffffffff81013689>] ? __switch_to+0x1f9/0x5c0
[670243.421267] [<ffffffff8109412a>] process_one_work+0x1aa/0x440
[670243.421269] [<ffffffff8109440b>] worker_thread+0x4b/0x4c0
[670243.421271] [<ffffffff810943c0>] ? process_one_work+0x440/0x440
[670243.421273] [<ffffffff810943c0>] ? process_one_work+0x440/0x440
[670243.421274] [<ffffffff8109a7c8>] kthread+0xd8/0xf0
[670243.421276] [<ffffffff8109a6f0>] ? kthread_create_on_node+0x1f0/0x1f0
[670243.421277] [<ffffffff817efe1f>] ret_from_fork+0x3f/0x70
[670243.421279] [<ffffffff8109a6f0>] ? kthread_create_on_node+0x1f0/0x1f0
[670243.421280] ---[ end trace 5cded7a882dfd5d1 ]---
[670243.421282] ceph: fill_inode badness ffff88179e2d9f28 10000004e91.fffffffffffffffe
this problem persisted through a reboot, and there is no fsck to help me.
I also tried with ceph-fuse, but it crashes when I access the file.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com