Re: ceph fuse closing stale session while still operable

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

i switched now debugging to ms = 10

when starting the dd i can see in the logs of osd:

2016-01-26 00:47:16.530046 7f086f404700  1 -- 10.0.0.1:6806/49658 >> :/0
pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=0 c=0x1dc2e9e0).accept
sd=292 10.0.0.91:56814/0
2016-01-26 00:47:16.530591 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=0
c=0x1dc2e9e0).accept peer addr is 10.0.0.91:0/1532
2016-01-26 00:47:16.530709 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=1
c=0x1dc2e9e0).accept of host_type 8, policy.lossy=1 policy.server=1
policy.standby=0 policy.resetcheck=0
2016-01-26 00:47:16.530724 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=1
c=0x1dc2e9e0).accept my proto 24, their proto 24
2016-01-26 00:47:16.530864 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=1
c=0x1dc2e9e0).accept:  setting up session_security.
2016-01-26 00:47:16.530877 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=0 pgs=0 cs=0 l=1
c=0x1dc2e9e0).accept new session
2016-01-26 00:47:16.530884 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).accept success, connect_seq = 1, sending READY
2016-01-26 00:47:16.530889 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).accept features 37154696925806591
2016-01-26 00:47:16.530912 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).register_pipe
2016-01-26 00:47:16.530935 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).discard_requeued_up_to 0
2016-01-26 00:47:16.530964 7f086f404700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).accept starting writer, state open
2016-01-26 00:47:16.531245 7f086f303700 10 -- 10.0.0.1:6806/49658 >>
10.0.0.91:0/1532 pipe(0x1f830000 sd=292 :6806 s=2 pgs=11 cs=1 l=1
c=0x1dc2e9e0).writer: state = open policy.server=1


And thats it, nothing more moves. And the dd is stuck.

[root@cn201 ~]# ceph daemon /var/run/ceph/ceph-client.admin.asok
objecter_requests
{
    "ops": [
        {
            "tid": 13,
            "pg": "6.ec7cd164",
            "osd": 1,
            "object_id": "10000000bdf.00000008",
            "object_locator": "@6",
            "target_object_id": "10000000bdf.00000008",
            "target_object_locator": "@6",
            "paused": 0,
            "used_replica": 0,
            "precalc_pgid": 0,
            "last_sent": "2016-01-26 00:47:28.783581",
            "attempts": 1,
            "snapid": "head",
            "snap_context": "1=[]",
            "mtime": "2016-01-26 00:47:28.769145",
            "osd_ops": [
                "write 0~4194304"
            ]
        },
        {
            "tid": 23,
            "pg": "6.1d0c5182",
            "osd": 1,
            "object_id": "10000000bdf.00000012",
            "object_locator": "@6",
            "target_object_id": "10000000bdf.00000012",
            "target_object_locator": "@6",
            "paused": 0,
            "used_replica": 0,
            "precalc_pgid": 0,
            "last_sent": "2016-01-26 00:47:28.917678",
            "attempts": 1,
            "snapid": "head",
            "snap_context": "1=[]",
            "mtime": "2016-01-26 00:47:28.904582",
            "osd_ops": [
                "write 0~4194304"
            ]
        },
        {
            "tid": 25,
            "pg": "6.97271370",
            "osd": 2,
            "object_id": "10000000bdf.00000014",
            "object_locator": "@6",
            "target_object_id": "10000000bdf.00000014",
            "target_object_locator": "@6",
            "paused": 0,
            "used_replica": 0,
            "precalc_pgid": 0,
            "last_sent": "2016-01-26 00:47:28.943763",
            "attempts": 1,
            "snapid": "head",
            "snap_context": "1=[]",
            "mtime": "2016-01-26 00:47:28.929857",
            "osd_ops": [
                "write 0~4194304"
            ]
        },

[....]


On another osd's i can see:

2016-01-26 00:11:36.224232 7f9c18596700  0 -- 10.0.0.1:6804/51935 >>
10.0.0.91:0/1536 pipe(0x1f955800 sd=276 :6804 s=0 pgs=0 cs=0 l=1
c=0x20379080).accept replacing existing (lossy) channel (new one lossy=1)
2016-01-26 00:26:36.312611 7f9c1a2b3700  0 -- 10.0.0.1:6804/51935 >>
10.0.0.91:0/1536 pipe(0x1f792000 sd=246 :6804 s=0 pgs=0 cs=0 l=1
c=0x155d02c0).accept replacing existing (lossy) channel (new one lossy=1)
2016-01-26 00:41:36.403063 7f9c1795c700  0 -- 10.0.0.1:6804/51935 >>
10.0.0.91:0/1536 pipe(0x20efa000 sd=254 :6804 s=0 pgs=0 cs=0 l=1
c=0x21bc1760).accept replacing existing (lossy) channel (new one lossy=1)


So for me, this is some kind of bug.

Without loaded rbd/ceph kernel module ceph-fuse will not work.

-- 
Mit freundlichen Gruessen / Best regards

Oliver Dzombic
IP-Interactive

mailto:info@xxxxxxxxxxxxxxxxx

Anschrift:

IP Interactive UG ( haftungsbeschraenkt )
Zum Sonnenberg 1-3
63571 Gelnhausen

HRB 93402 beim Amtsgericht Hanau
Geschäftsführung: Oliver Dzombic

Steuer Nr.: 35 236 3622 1
UST ID: DE274086107
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux