Re: Fwd: mds port 6800 socket closed when accessing mount point

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sage,

On Wed, Oct 27, 2010 at 8:16 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
> Hi Yonggang,
>
> Sorry, I misread the debug output.  You're right, the pid's are off by one
> because the daemons do a fork() and exit, and it's the child that survives
> and stays active.
Thank you for explaining this to me. That makes sense. So what can I
do to make it working correctly, or is this how it works?
>
> In any case, the source of the offending message is
>
> 192.168.89.131:0/753148386
>
> The large number at the end is just random.  I'm guessing that's the
> kernel client mounting/trying to mount the file system.  What kernel
> version, or version of the kernel module are you using?
>
Here's the kernel version of all my VMs:
mds0:~# uname -a
Linux mds0 2.6.32-5-686 #1 SMP Tue Oct 19 14:40:34 UTC 2010 i686 GNU/Linux

The mount command I was using is:
mount -t ceph mds0:/ /ceph
Do you mean the failure is due to the client kernel?

> Thanks-
> sage
>

Thank you!

>
> On Wed, 27 Oct 2010, Yonggang Liu wrote:
>
>> ---------- Forwarded message ----------
>> From: Yonggang Liu <myidpt@xxxxxxxxx>
>> Date: Wed, Oct 27, 2010 at 7:32 PM
>> Subject: Re: mds port 6800 socket closed when accessing mount point
>> To: Sage Weil <sage@xxxxxxxxxxxx>
>>
>>
>> Hi Sage,
>>
>> Thank you very much for your reply. In your question, the PID 1978 is
>> not a running process; it is the PID Ceph considers the mds0 process
>> is using. But the real PID of mds0 was not that one (seem to be 1979).
>> I redid the process and gathered the results. I think I see a little
>> bit of the problem now. When I'm starting Ceph, the program prints out
>> the PIDs of the daemons running on each machine, but these PIDs are
>> not the exact PIDs of the real running daemons.
>>
>> I started the system and got the PID information:
>> mds0:~# /etc/init.d/ceph -a start
>> === mon.0 ===
>> Starting Ceph mon0 on mds0...
>>  ** WARNING: Ceph is still under heavy development, and is only suitable for **
>>  **          testing and review.  Do not trust it with important data.       **
>> starting mon.0 rank 0 at 192.168.89.133:6789/0 mon_data /data/mon0
>> fsid 9ff93f53-5e63-011c-4437-8e3f780cfcde
>> === mds.0 ===
>> Starting Ceph mds0 on mds0...
>>  ** WARNING: Ceph is still under heavy development, and is only suitable for **
>>  **          testing and review.  Do not trust it with important data.       **
>> starting mds.0 at 0.0.0.0:6800/1824
>> === osd.0 ===
>> Starting Ceph osd0 on osd0...
>>  ** WARNING: Ceph is still under heavy development, and is only suitable for **
>>  **          testing and review.  Do not trust it with important data.       **
>> starting osd0 at 0.0.0.0:6800/1622 osd_data /data/osd0 /data/osd0/journal
>> === osd.1 ===
>> Starting Ceph osd1 on osd1...
>>  ** WARNING: Ceph is still under heavy development, and is only suitable for **
>>  **          testing and review.  Do not trust it with important data.       **
>> starting osd1 at 0.0.0.0:6800/1616 osd_data /data/osd1 /data/osd1/journal
>>
>> So from the print, the PIDs known by Ceph are:
>> moon.0       0
>> mds.0         1824
>> osd.0          1622
>> osd.1          1616
>>
>> But I checked the PIDs by using "ps uax | grep ceph" on the 3
>> machines, I got this:
>> On mds0:
>> root      1792  1.0  1.2  46412  3260 ?        S<sl 20:14   0:07
>> /usr/bin/cmon -i 0 -c /etc/ceph/ceph.conf
>> root      1825 28.3  1.2  72228  3076 ?        S<sl 20:14   3:06
>> /usr/bin/cmds -i 0 -c /etc/ceph/ceph.conf
>> On osd0:
>> root      1623  0.9  8.3 214820 21356 ?        S<sl 20:07   0:01
>> /usr/bin/cosd -i 0 -c /tmp/ceph.conf.1747
>> On osd1:
>> root      1617  0.7  7.5 213780 19264 ?        S<sl 20:07   0:01
>> /usr/bin/cosd -i 1 -c /tmp/ceph.conf.1747
>> (Note the second fields are the PIDs)
>>
>> Run "netstat -anp | grep 6800" on mds0, I got:
>> mds0:~# netstat -anp | grep 6800
>> tcp        0      0 0.0.0.0:6800            0.0.0.0:*
>> LISTEN      1825/cmds
>> tcp        0      0 192.168.89.133:33216    192.168.89.135:6800
>> ESTABLISHED 1825/cmds
>> tcp        0      0 192.168.89.133:56908    192.168.89.134:6800
>> ESTABLISHED 1825/cmds
>> tcp        0      0 192.168.89.133:6800     192.168.89.131:45334
>> ESTABLISHED 1825/cmds
>> (I can see PID 1825 is taking care of the client - mds0 communication)
>>
>> At the end of file "mds.0.log", I saw the PID 1824:
>> 2010-10-27 20:57:45.537898 b35e4b70 failed to decode message of type
>> 784 v2: buffer::end_of_buffer
>> 2010-10-27 20:57:45.538652 b35e4b70 -- 192.168.89.133:6800/1824 >>
>> 192.168.89.131:0/753148386 pipe(0x9f43488 sd=-1 pgs=9 cs=7 l=0).fault
>> with nothing to send, going to standby
>> 2010-10-27 20:58:18.044950 b30fdb70 -- 192.168.89.133:6800/1824 >>
>> 192.168.89.131:0/753148386 pipe(0x9f7cca8 sd=10 pgs=0 cs=0 l=0).accept
>> peer addr is really 192.168.89.131:0/753148386 (socket is
>> 192.168.89.131:45342/0)
>> 2010-10-27 20:58:18.045247 b30fdb70 -- 192.168.89.133:6800/1824 >>
>> 192.168.89.131:0/753148386 pipe(0x9f7cca8 sd=10 pgs=0 cs=0 l=0).accept
>> connect_seq 7 vs existing 7 state 3
>> 2010-10-27 20:58:18.053628 b30fdb70 failed to decode message of type
>> 784 v2: buffer::end_of_buffer
>> 2010-10-27 20:58:18.054824 b30fdb70 -- 192.168.89.133:6800/1824 >>
>> 192.168.89.131:0/753148386 pipe(0x9f7cca8 sd=-1 pgs=10 cs=8 l=0).fault
>> initiating reconnect
>>
>> So this is strange. It seems the PIDs known by Ceph are not the real
>> PIDs of the processes. I don't know how this could happen.
>>
>> The version of the Ceph code is:
>> mds0:~# ceph --version
>> ceph version 0.22.1 (commit:7464f9688001aa89f9673ba14e6d075d0ee33541)
>>
>> Thank you very much!
>>
>>
>> On Wed, Oct 27, 2010 at 4:47 PM, Sage Weil <sage@xxxxxxxxxxxx> wrote:
>> > Hi,
>> >
>> > On Wed, 27 Oct 2010, Yonggang Liu wrote:
>> >> 2010-10-27 15:56:23.332473 b352eb70 failed to decode message of type
>> >> 784 v4865: buffer::end_of_buffer
>> >
>> > This is the problem.  Can you tell what 192.168.89.133:6800/1978 is?
>> > (That 1978 is a pid, btw.)  It's probably a 'cfuse' I'm guessing?  (The
>> > ceph daemons all start with 'c', but don't have 'ceph' in the name.)
>> >
>> > 784 is a client_caps message. The 'v4865' looks like an uninitialized
>> > variable to me, as it should be a small integer.
>> >
>> > Whatever daemon that is, can you run it with the '-v' argument to see
>> > exactly which version it is?
>> >
>> > Thanks!
>> > sage
>> >
>> >
>> >
>> >
>> >> 2010-10-27 15:56:23.332730 b352eb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd0168 sd=-1 pgs=10 cs=8
>> >> l=0).fault with nothing to send, going to standby
>> >> 2010-10-27 15:56:52.022573 b332cb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd1720 sd=12 pgs=0 cs=0
>> >> l=0).accept peer addr is really 192.168.89.131:0/1344076202 (socket is
>> >> 192.168.89.131:41726/0)
>> >> 2010-10-27 15:56:52.022945 b332cb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd1720 sd=12 pgs=0 cs=0
>> >> l=0).accept connect_seq 8 vs existing 8 state 3
>> >> 2010-10-27 15:56:52.025299 b332cb70 failed to decode message of type
>> >> 784 v4865: buffer::end_of_buffer
>> >> 2010-10-27 15:56:52.025569 b332cb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd1720 sd=-1 pgs=11 cs=9
>> >> l=0).fault with nothing to send, going to standby
>> >> 2010-10-27 15:57:48.106337 b362fb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd07c0 sd=12 pgs=0 cs=0
>> >> l=0).accept peer addr is really 192.168.89.131:0/1344076202 (socket is
>> >> 192.168.89.131:41727/0)
>> >> 2010-10-27 15:57:48.106522 b362fb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd07c0 sd=12 pgs=0 cs=0
>> >> l=0).accept connect_seq 9 vs existing 9 state 3
>> >> 2010-10-27 15:57:48.109498 b362fb70 failed to decode message of type
>> >> 784 v4865: buffer::end_of_buffer
>> >> 2010-10-27 15:57:48.109761 b362fb70 -- 192.168.89.133:6800/1978 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x9fd07c0 sd=-1 pgs=12 cs=10
>> >> l=0).fault with nothing to send, going to standby
>> >>
>> >> On osd0
>> >> osd.0.log:
>> >> 2010-10-27 15:55:01.470502 --- 1633 opened log /var/log/ceph/osd.0.log ---
>> >> ceph version 0.22.1 (commit:7464f9688001aa89f9673ba14e6d075d0ee33541)
>> >> 2010-10-27 15:55:01.485828 b72d38e0 filestore(/data/osd0) mkfs in /data/osd0
>> >> 2010-10-27 15:55:01.486106 b72d38e0 filestore(/data/osd0) mkfs
>> >> removing old file fsid
>> >> 2010-10-27 15:55:01.516527 b72d38e0 filestore(/data/osd0) mkjournal
>> >> created journal on /journal
>> >> 2010-10-27 15:55:01.516734 b72d38e0 filestore(/data/osd0) mkfs done in
>> >> /data/osd0
>> >> 2010-10-27 15:55:01.519606 b72d38e0 filestore(/data/osd0) mount did
>> >> NOT detect btrfs
>> >> 2010-10-27 15:55:01.519794 b72d38e0 filestore(/data/osd0) mount found snaps <>
>> >> 2010-10-27 15:55:01.548307 b5acfb70 FileStore::op_tp worker finish
>> >> 2010-10-27 15:55:01.548455 b52ceb70 FileStore::op_tp worker finish
>> >> 2010-10-27 15:55:01.548812 b72d38e0 journal close /journal
>> >> 2010-10-27 15:55:25.201288 --- 1656 opened log /var/log/ceph/osd.0.log ---
>> >> ceph version 0.22.1 (commit:7464f9688001aa89f9673ba14e6d075d0ee33541)
>> >> 2010-10-27 15:55:25.223381 b74508e0 filestore(/data/osd0) mount did
>> >> NOT detect btrfs
>> >> 2010-10-27 15:55:25.224253 b74508e0 filestore(/data/osd0) mount found snaps <>
>> >> 2010-10-27 15:55:25.225396 b74508e0 journal read_entry 4096 : seq 1 203 bytes
>> >> 2010-10-27 15:55:26.841749 abffeb70 -- 0.0.0.0:6801/1656 >>
>> >> 192.168.89.135:6801/1675 pipe(0x94ced18 sd=13 pgs=0 cs=0 l=0).connect
>> >> claims to be 0.0.0.0:6801/1675 not 192.168.89.135:6801/1675 -
>> >> presumably this is the same node!
>> >> 2010-10-27 15:56:04.219613 abaf9b70 -- 192.168.89.134:6800/1656 >>
>> >> 192.168.89.131:0/1344076202 pipe(0x96463e0 sd=16 pgs=0 cs=0
>> >> l=0).accept peer addr is really 192.168.89.131:0/1344076202 (socket is
>> >> 192.168.89.131:36746/0)
>> >>
>> >> On osd1
>> >> osd.1.log:
>> >> 2010-10-27 15:54:59.752615 --- 1652 opened log /var/log/ceph/osd.1.log ---
>> >> ceph version 0.22.1 (commit:7464f9688001aa89f9673ba14e6d075d0ee33541)
>> >> 2010-10-27 15:54:59.766128 b73518e0 filestore(/data/osd1) mkfs in /data/osd1
>> >> 2010-10-27 15:54:59.766658 b73518e0 filestore(/data/osd1) mkfs
>> >> removing old file fsid
>> >> 2010-10-27 15:54:59.796938 b73518e0 filestore(/data/osd1) mkjournal
>> >> created journal on /journal
>> >> 2010-10-27 15:54:59.797816 b73518e0 filestore(/data/osd1) mkfs done in
>> >> /data/osd1
>> >> 2010-10-27 15:54:59.800957 b73518e0 filestore(/data/osd1) mount did
>> >> NOT detect btrfs
>> >> 2010-10-27 15:54:59.801087 b73518e0 filestore(/data/osd1) mount found snaps <>
>> >> 2010-10-27 15:54:59.832202 b534cb70 FileStore::op_tp worker finish
>> >> 2010-10-27 15:54:59.832504 b5b4db70 FileStore::op_tp worker finish
>> >> 2010-10-27 15:54:59.832723 b73518e0 journal close /journal
>> >> 2010-10-27 15:55:23.050042 --- 1675 opened log /var/log/ceph/osd.1.log ---
>> >> ceph version 0.22.1 (commit:7464f9688001aa89f9673ba14e6d075d0ee33541)
>> >> 2010-10-27 15:55:23.056671 b72e18e0 filestore(/data/osd1) mount did
>> >> NOT detect btrfs
>> >> 2010-10-27 15:55:23.056921 b72e18e0 filestore(/data/osd1) mount found snaps <>
>> >> 2010-10-27 15:55:23.057368 b72e18e0 journal read_entry 4096 : seq 1 203 bytes
>> >> 2010-10-27 15:55:23.207216 b12d4b70 osd1 2 map says i am down or have
>> >> a different address.  switching to boot state.
>> >> 2010-10-27 15:55:23.207540 b12d4b70 log [WRN] : map e2 wrongly marked me down
>> >>
>> >> I noticed on osd1, the last two lines of osd.1.log are unusual, but
>> >> I'm not sure if they are the reason to the problem ...
>> >>
>> >> Thank you very much,
>> >>
>> >>
>> >> On Wed, Oct 27, 2010 at 1:10 PM, Colin McCabe <cmccabe@xxxxxxxxxxxxxx> wrote:
>> >> > Hi Yonggang,
>> >> >
>> >> > Are all of the daemons still running? What is at the end of the logfiles?
>> >> >
>> >> > regards,
>> >> > Colin
>> >> >
>> >> >
>> >> > On Wed, Oct 27, 2010 at 9:42 AM, Yonggang Liu <myidpt@xxxxxxxxx> wrote:
>> >> >> Hello,
>> >> >>
>> >> >> I'm totally new to Ceph. Last a few days I set up 4 VMs to run Ceph:
>> >> >> "mds0" for the metadata server and monitor, "osd0" and "osd1" for two
>> >> >> data servers, and "client" for the client machine. The VMs are running
>> >> >> Debian 5.0 with kernel 2.6.32-5-686 (Ceph module is enabled).
>> >> >> I followed "Building kernel client" and "Debian" from the wiki, and I
>> >> >> was able to start Ceph and mount Ceph at the client. But the problem
>> >> >> is, the mounted point always fail with an infinite response time
>> >> >> (after I mount Ceph for about 1 min or less). To illustrate it better,
>> >> >> I will show you the information I got on the client and mds0 machines:
>> >> >>
>> >> >> mds0 (192.168.89.133):
>> >> >> debian:~# mkcephfs -c /etc/ceph/ceph.conf --allhosts -v
>> >> >> (A lot of info)
>> >> >> debian:~# /etc/init.d/ceph -a start
>> >> >> (some info)
>> >> >>
>> >> >> client (192.168.89.131):
>> >> >> debian:~# mount -t ceph 192.168.89.133:/ /ceph
>> >> >> debian:~# cd /ceph
>> >> >> debian:/ceph# cp ~/app_ch.xls .
>> >> >> debian:/ceph# ls
>> >> >> (waiting for ever)
>> >> >> ^C
>> >> >>
>> >> >> After the failure I ran dmesg at the client side and got:
>> >> >> client (192.168.89.131):
>> >> >> debian:/ceph# dmesg -c
>> >> >> [  636.664425] ceph: loaded (mon/mds/osd proto 15/32/24, osdmap 5/5 5/5)
>> >> >> [  636.694973] ceph: client4100 fsid 423ad64c-bbf0-3011-bb47-36a89f8787c6
>> >> >> [  636.700716] ceph: mon0 192.168.89.133:6789 session established
>> >> >> [  664.114551] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  664.848722] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  665.914923] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  667.840396] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  672.054106] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  680.894531] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  696.928496] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  720.171754] ceph: mds0 caps stale
>> >> >> [  728.999701] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >> [  794.640943] ceph: mds0 192.168.89.133:6800 socket closed
>> >> >>
>> >> >> Immediately after the failure, I ran netstat at mds0:
>> >> >> mds0 (192.168.89.133):
>> >> >> debian:~# netstat -anp
>> >> >> Active Internet connections (servers and established)
>> >> >> Proto Recv-Q Send-Q Local Address           Foreign Address
>> >> >> State       PID/Program name
>> >> >> tcp        0      0 0.0.0.0:6800            0.0.0.0:*
>> >> >> LISTEN      1889/cmds
>> >> >> tcp        0      0 0.0.0.0:22              0.0.0.0:*
>> >> >> LISTEN      1529/sshd
>> >> >> tcp        0      0 192.168.89.133:6789     0.0.0.0:*
>> >> >> LISTEN      1840/cmon
>> >> >> tcp        0      0 192.168.89.133:6789     192.168.89.131:56855
>> >> >> ESTABLISHED 1840/cmon
>> >> >> tcp        0      0 192.168.89.133:43647    192.168.89.133:6789
>> >> >> ESTABLISHED 1889/cmds
>> >> >> tcp        0      0 192.168.89.133:22       192.168.89.1:58304
>> >> >> ESTABLISHED 1530/0
>> >> >> tcp        0      0 192.168.89.133:39826    192.168.89.134:6800
>> >> >> ESTABLISHED 1889/cmds
>> >> >> tcp        0      0 192.168.89.133:6789     192.168.89.134:41289
>> >> >> ESTABLISHED 1840/cmon
>> >> >> tcp        0      0 192.168.89.133:6800     192.168.89.131:52814
>> >> >> TIME_WAIT   -
>> >> >> tcp        0      0 192.168.89.133:6789     192.168.89.135:41021
>> >> >> ESTABLISHED 1840/cmon
>> >> >> tcp        0      0 192.168.89.133:42069    192.168.89.135:6800
>> >> >> ESTABLISHED 1889/cmds
>> >> >> tcp        0      0 192.168.89.133:6789     192.168.89.133:43647
>> >> >> ESTABLISHED 1840/cmon
>> >> >> tcp        0      0 192.168.89.133:6800     192.168.89.131:52815
>> >> >> TIME_WAIT   -
>> >> >> tcp        0      0 192.168.89.133:6800     192.168.89.131:52816
>> >> >> TIME_WAIT   -
>> >> >> tcp6       0      0 :::22                   :::*
>> >> >> LISTEN      1529/sshd
>> >> >> udp        0      0 0.0.0.0:68              0.0.0.0:*
>> >> >>         1490/dhclient3
>> >> >> Active UNIX domain sockets (servers and established)
>> >> >> Proto RefCnt Flags       Type       State         I-Node   PID/Program
>> >> >> name    Path
>> >> >> unix  2      [ ]         DGRAM                    2972     546/udevd
>> >> >>        @/org/kernel/udev/udevd
>> >> >> unix  4      [ ]         DGRAM                    5343
>> >> >> 1358/rsyslogd       /dev/log
>> >> >> unix  2      [ ]         DGRAM                    5662     1530/0
>> >> >> unix  2      [ ]         DGRAM                    5486     1490/dhclient3
>> >> >> debian:~#
>> >> >> debian:~# dmesg -c
>> >> >> debian:~# (nothing shows up)
>> >> >>
>> >> >> I saw the port 6800 on the metadata server talking with the client is
>> >> >> on "TIME_WAIT" stage. That means the connection is closed.
>> >> >> This is the ceph.conf I have:
>> >> >> [global]
>> >> >>       pid file = /var/run/ceph/$type.$id.pid
>> >> >> [mon]
>> >> >>       mon data = /data/mon$id
>> >> >>       mon subscribe interval = 6000
>> >> >>       mon osd down out interval = 6000
>> >> >> [mon0]
>> >> >>       host = mds0
>> >> >>       mon addr = 192.168.89.133:6789
>> >> >> [mds]
>> >> >>       mds session timeout = 6000
>> >> >>       mds session autoclose = 6000
>> >> >>       mds client lease = 6000
>> >> >>       keyring = /data/keyring.$name
>> >> >> [mds0]
>> >> >>       host = mds0
>> >> >> [osd]
>> >> >>       sudo = true
>> >> >>       osd data = /data/osd$id
>> >> >>       osd journal = /journal
>> >> >>       osd journal size = 1024
>> >> >>       filestore journal writeahead = true
>> >> >> [osd0]
>> >> >>       host = osd0
>> >> >> [osd1]
>> >> >>       host = osd1
>> >> >> [group everyone]
>> >> >>       addr = 0.0.0.0/0
>> >> >> [mount]
>> >> >>       allow = %everyone
>> >> >> ;-----------------------------------end-----------------------------------
>> >> >>
>> >> >> The Ceph version I was using is 0.22.1.
>> >> >>
>> >> >> Can anyone help me to solve this problem? Thanks in advance!
>> >> >> --
>> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >>
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> Yonggang Liu
>> >> Advanced Computing and Information Systems Laboratory
>> >> University of Florida
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >>
>> >>
>>
>>
>>
>> --
>> Yonggang Liu
>> Advanced Computing and Information Systems Laboratory
>> University of Florida
>>
>>
>>
>> --
>> Yonggang Liu
>> Advanced Computing and Information Systems Laboratory
>> University of Florida
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>



-- 
Yonggang Liu
Advanced Computing and Information Systems Laboratory
University of Florida
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux