Re: keyring generation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




(2014/02/05 23:49), Alfredo Deza wrote:
On Mon, Feb 3, 2014 at 11:28 AM, Kei.masumoto <kei.masumoto@xxxxxxxxx> wrote:
Hi Alfredo,

Thanks for your reply!

I think I pasted all logs from ceph.log,  but anyway, I re-executed
"ceph-deploy mon create-initial again"
Does that make sense? It seems like stack strace are added...
Those seem bad enough. There is a ticket open for these type of
tracebacks that should be gone with the
up coming release of ceph-deploy.

Your monitor does seem like in a good state. Have you checked the
monitor logs to see if they are complaining
about something?

I would also raise the log level in ceph.conf for the monitors specifically to:

   debug mon = 10


Thanks for your reply.
I did "debug mon = 10", but I could not find any error logs in /var/log/ceph/ceph-mon.ceph1.log. So I tried to let ceph-create-keys generate logs to files, and inserted log by myself for debugging purpose. Then I found get_key() in ceph-create-keys complains like below. ( first line is inserted by me)

> INFO:ceph-create-keys: ceph --cluster=ceph --name=mon. --keyring=/var/lib/ceph/mon/ceph-ceph1/keyring auth get-or-create client.admin mon allow * osd allow * mds allow
> INFO:ceph-create-keys:Talking to monitor...
> INFO:ceph-create-keys:Cannot get or create admin key, permission denied

So I tried,
> root@ceph1:~/my-cluster# chown -R ceph:ceph /var/lib/ceph/mon/
> root@ceph1:~/my-cluster# start ceph-all && stop ceph-all
> ceph-all start/running
> ceph-all stop/waiting

Then I re-tried:
> ceph@ceph1:~/my-cluster$ ceph-deploy --overwrite-conf mon create-initial

After that, I found some files are still owned b root. Is this a correct behavior?
> root@ceph1:~/my-cluster# ls -l /var/lib/ceph/mon/ceph-ceph1/store.db
> total 1184
> -rw-r--r-- 1 ceph ceph 1081168 Feb  8 02:25 000133.sst
> -rw-r--r-- 1 ceph ceph   25530 Feb  8 02:38 000135.sst
> -rw-r--r-- 1 ceph ceph   25530 Feb  8 02:38 000138.sst
> -rw-r--r-- 1 root root   25530 Feb  8 02:44 000141.sst
> -rw-r--r-- 1 root root   65536 Feb  8 02:44 000142.log
> -rw-r--r-- 1 root root      16 Feb  8 02:44 CURRENT
> -rw-r--r-- 1 ceph ceph       0 Jan 26 05:50 LOCK
> -rw-r--r-- 1 ceph ceph     315 Jan 26 06:28 LOG
> -rw-r--r-- 1 ceph ceph      57 Jan 26 05:50 LOG.old
> -rw-r--r-- 1 root root   65536 Feb  8 02:44 MANIFEST-000140

Also, looks like ceph-create-keys and ceph-mon are launched by root. Is this correct?
> root@ceph1:~/my-cluster# ps -ef | grep ceph
> root 18943 1 0 03:11 ? 00:00:00 /bin/sh -e -c /usr/bin/ceph-mon --cluster="${cluster:-ceph}" -i "$id" -f /bin/sh > root 18944 18943 0 03:11 ? 00:00:00 /usr/bin/ceph-mon --cluster=ceph -i ceph1 -f > root 18945 1 0 03:11 ? 00:00:00 /bin/sh -e -c /usr/sbin/ceph-create-keys --cluster="${cluster:-ceph}" -i "${id:-$(hostname)}" /bin/sh > root 18946 18945 0 03:11 ? 00:00:00 /usr/bin/python /usr/sbin/ceph-create-keys --cluster=ceph -i ceph1 > root 18960 18946 1 03:11 ? 00:00:00 /usr/bin/python /usr/bin/ceph --cluster=ceph --name=mon. --keyring=/var/lib/ceph/mon/ceph-ceph1/keyring auth get-or-create client.admin mon allow * osd allow * mds allow > root 19036 16175 0 03:11 pts/7 00:00:00 tail -f /var/log/ceph/ceph-mon.ceph1.log




--------------------------------------------------------------------------------------------------------------------------

[ceph_deploy.cli][INFO  ] Invoked (1.3.4): /usr/bin/ceph-deploy mon
create-initial
[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts ceph1
[ceph_deploy.mon][DEBUG ] detecting platform for host ceph1 ...
[ceph1][DEBUG ] connected to host: ceph1
[ceph1][DEBUG ] detect platform information from remote host
[ceph1][DEBUG ] detect machine type
[ceph_deploy.mon][INFO  ] distro info: Ubuntu 13.04 raring
[ceph1][DEBUG ] determining if provided host has same hostname in remote
[ceph1][DEBUG ] get remote short hostname
[ceph1][DEBUG ] deploying mon to ceph1
[ceph1][DEBUG ] get remote short hostname
[ceph1][DEBUG ] remote hostname: ceph1
[ceph1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph1][DEBUG ] create the mon path if it does not exist
[ceph1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph1/done
[ceph1][DEBUG ] create a done file to avoid re-doing the mon deployment
[ceph1][DEBUG ] create the init path if it does not exist
[ceph1][DEBUG ] locating the `service` executable...
[ceph1][INFO  ] Running command: sudo initctl emit ceph-mon cluster=ceph
id=ceph1
[ceph1][INFO  ] Running command: sudo ceph --cluster=ceph --admin-daemon
/var/run/ceph/ceph-mon.ceph1.asok mon_status
[ceph1][DEBUG ]
********************************************************************************

[ceph1][DEBUG ] status for monitor: mon.ceph1
[ceph1][DEBUG ] {
[ceph1][DEBUG ]   "election_epoch": 1,
[ceph1][DEBUG ]   "extra_probe_peers": [],
[ceph1][DEBUG ]   "monmap": {

[ceph1][DEBUG ]     "created": "0.000000",
[ceph1][DEBUG ]     "epoch": 1,

[ceph1][DEBUG ]     "fsid": "26835656-6b29-455d-9d1f-545cad8f1e23",
[ceph1][DEBUG ]     "modified": "0.000000",
[ceph1][DEBUG ]     "mons": [
[ceph1][DEBUG ]       {

[ceph1][DEBUG ]         "addr": "192.168.111.11:6789/0",
[ceph1][DEBUG ]         "name": "ceph1",
[ceph1][DEBUG ]         "rank": 0
[ceph1][DEBUG ]       }
[ceph1][DEBUG ]     ]
[ceph1][DEBUG ]   },

[ceph1][DEBUG ]   "name": "ceph1",
[ceph1][DEBUG ]   "outside_quorum": [],
[ceph1][DEBUG ]   "quorum": [
[ceph1][DEBUG ]     0
[ceph1][DEBUG ]   ],
[ceph1][DEBUG ]   "rank": 0,

[ceph1][DEBUG ]   "state": "leader",
[ceph1][DEBUG ]   "sync_provider": []
[ceph1][DEBUG ] }
[ceph1][DEBUG ]
********************************************************************************

[ceph1][INFO  ] monitor: mon.ceph1 is running
[ceph1][INFO  ] Running command: sudo ceph --cluster=ceph --admin-daemon
/var/run/ceph/ceph-mon.ceph1.asok mon_status
[ceph_deploy.mon][INFO  ] processing monitor mon.ceph1
[ceph1][DEBUG ] connected to host: ceph1
[ceph1][INFO  ] Running command: sudo ceph --cluster=ceph --admin-daemon
/var/run/ceph/ceph-mon.ceph1.asok mon_status
[ceph_deploy.mon][INFO  ] mon.ceph1 monitor has reached quorum!
[ceph_deploy.mon][INFO  ] all initial monitors are running and have formed
quorum
[ceph_deploy.mon][INFO  ] Running gatherkeys...
gatherkeys.fetch_file    Namespace(cluster='ceph', dry_run=False,
func=<function mon at 0xe14e60>, mon=['ceph1'], overwrite_conf=False,
prog='ceph-deploy', quiet=False, subcommand='create-initial', username=None,
verbose=False) :: /etc/ceph/ceph.client.admin.keyring ::
ceph.client.admin.keyring :: ['ceph1']

[ceph_deploy.gatherkeys][DEBUG ] Checking ceph1 for
/etc/ceph/ceph.client.admin.keyring
[ceph1][DEBUG ] connected to host: ceph1
[ceph1][DEBUG ] detect platform information from remote host
[ceph1][DEBUG ] detect machine type
[ceph1][DEBUG ] fetch remote file
[ceph_deploy.gatherkeys][WARNIN] Unable to find
/etc/ceph/ceph.client.admin.keyring on ['ceph1']
Traceback (most recent call last):
   File "<string>", line 1, in <module>
   File "<string>", line 6, in <module>
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 1220, in serve
     gatherkeys.fetch_file    Namespace(cluster='ceph', dry_run=False,
func=<function mon at 0xe14e60>, mon=['ceph1'], overwrite_conf=False,
prog='ceph-deploy', quiet=False, subcommand='create-initial', username=None,
verbose=False) :: /var/lib/ceph/mon/ceph-{hostname}/keyring ::
ceph.mon.keyring :: ['ceph1']
[ceph_deploy.gatherkeys][DEBUG ] Have ceph.mon.keyring
gatherkeys.fetch_file    Namespace(cluster='ceph', dry_run=False,
func=<function mon at 0xe14e60>, mon=['ceph1'], overwrite_conf=False,
prog='ceph-deploy', quiet=False, subcommand='create-initial', username=None,
verbose=False) :: /var/lib/ceph/bootstrap-osd/ceph.keyring ::
ceph.bootstrap-osd.keyring :: ['ceph1']
SlaveGateway(io=io, id=id, _startcount=2).serve()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 764, in serve

     [ceph_deploy.gatherkeys][DEBUG ] Checking ceph1 for
/var/lib/ceph/bootstrap-osd/ceph.keyring
self._io.close_write()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 105, in close_write
     self.outfile.close()
IOError: close() called during concurrent operation on the same file object.

[ceph1][DEBUG ] connected to host: ceph1
[ceph1][DEBUG ] detect platform information from remote host
[ceph1][DEBUG ] detect machine type
[ceph1][DEBUG ] fetch remote file
[ceph_deploy.gatherkeys][WARNIN] Unable to find
/var/lib/ceph/bootstrap-osd/ceph.keyring on ['ceph1']
Traceback (most recent call last):
   File "<string>", line 1, in <module>
   File "<string>", line 6, in <module>
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 1220, in serve
     SlaveGateway(io=io, id=id, _startcount=2).serve()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 764, in serve
     self._io.close_write()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 105, in close_write
     gatherkeys.fetch_file    Namespace(cluster='ceph', dry_run=False,
func=<function mon at 0xe14e60>, mon=['ceph1'], overwrite_conf=False,
prog='ceph-deploy', quiet=False, subcommand='create-initial', username=None,
verbose=False) :: /var/lib/ceph/bootstrap-mds/ceph.keyring ::
ceph.bootstrap-mds.keyring :: ['ceph1']
self.outfile.close()
IOError: close() called during concurrent operation on the same file object.

[ceph_deploy.gatherkeys][DEBUG ] Checking ceph1 for
/var/lib/ceph/bootstrap-mds/ceph.keyring
[ceph1][DEBUG ] connected to host: ceph1
[ceph1][DEBUG ] detect platform information from remote host
[ceph1][DEBUG ] detect machine type
[ceph1][DEBUG ] fetch remote file
[ceph_deploy.gatherkeys][WARNIN] Unable to find
/var/lib/ceph/bootstrap-mds/ceph.keyring on ['ceph1']
Traceback (most recent call last):
   File "<string>", line 1, in <module>
   File "<string>", line 6, in <module>
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 1220, in serve
     SlaveGateway(io=io, id=id, _startcount=2).serve()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 764, in serve
     self._io.close_write()
   File
"/usr/lib/python2.7/dist-packages/ceph_deploy/lib/remoto/lib/execnet/gateway_base.py",
line 105, in close_write
     self.outfile.close()
IOError: close() called during concurrent operation on the same file object.

----------------------------------------------------------------------------------------------------------









(2014/02/04 1:12), Alfredo Deza wrote:
On Mon, Feb 3, 2014 at 10:07 AM, Kei.masumoto <kei.masumoto@xxxxxxxxx>
wrote:
Hi Alfredo,

Thanks for reply!  I pasted the logs below.


------------------------------------------------------------------------------------
2014-02-01 14:06:33,350 [ceph_deploy.cli][INFO  ] Invoked (1.3.4):
/usr/bin/ceph-deploy mon create-initial
2014-02-01 14:06:33,353 [ceph_deploy.mon][DEBUG ] Deploying mon, cluster
ceph hosts ceph1
2014-02-01 14:06:33,354 [ceph_deploy.mon][DEBUG ] detecting platform for
host ceph1 ...
2014-02-01 14:06:33,770 [ceph1][DEBUG ] connected to host: ceph1
2014-02-01 14:06:33,775 [ceph1][DEBUG ] detect platform information from
remote host
2014-02-01 14:06:33,874 [ceph1][DEBUG ] detect machine type
2014-02-01 14:06:33,909 [ceph_deploy.mon][INFO  ] distro info: Ubuntu
13.04
raring
2014-02-01 14:06:33,910 [ceph1][DEBUG ] determining if provided host has
same hostname in remote
2014-02-01 14:06:33,911 [ceph1][DEBUG ] get remote short hostname
2014-02-01 14:06:33,914 [ceph1][DEBUG ] deploying mon to ceph1
2014-02-01 14:06:33,915 [ceph1][DEBUG ] get remote short hostname
2014-02-01 14:06:33,917 [ceph1][DEBUG ] remote hostname: ceph1
2014-02-01 14:06:33,919 [ceph1][DEBUG ] write cluster configuration to
/etc/ceph/{cluster}.conf
2014-02-01 14:06:33,933 [ceph1][DEBUG ] create the mon path if it does
not
exist
2014-02-01 14:06:33,939 [ceph1][DEBUG ] checking for done path:
/var/lib/ceph/mon/ceph-ceph1/done
2014-02-01 14:06:33,941 [ceph1][DEBUG ] create a done file to avoid
re-doing
the mon deployment
2014-02-01 14:06:33,944 [ceph1][DEBUG ] create the init path if it does
not
exist
2014-02-01 14:06:33,946 [ceph1][DEBUG ] locating the `service`
executable...
2014-02-01 14:06:33,949 [ceph1][INFO  ] Running command: sudo initctl
emit
ceph-mon cluster=ceph id=ceph1
2014-02-01 14:06:36,119 [ceph1][INFO  ] Running command: sudo ceph
--cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph1.asok
mon_status
2014-02-01 14:06:36,805 [ceph1][DEBUG ]

********************************************************************************
2014-02-01 14:06:36,807 [ceph1][DEBUG ] status for monitor: mon.ceph1
2014-02-01 14:06:36,809 [ceph1][DEBUG ] {
2014-02-01 14:06:36,810 [ceph1][DEBUG ]   "election_epoch": 1,
2014-02-01 14:06:36,812 [ceph1][DEBUG ]   "extra_probe_peers": [],
2014-02-01 14:06:36,813 [ceph1][DEBUG ]   "monmap": {
2014-02-01 14:06:36,814 [ceph1][DEBUG ]     "created": "0.000000",
2014-02-01 14:06:36,815 [ceph1][DEBUG ]     "epoch": 1,
2014-02-01 14:06:36,815 [ceph1][DEBUG ]     "fsid":
"26835656-6b29-455d-9d1f-545cad8f1e23",
2014-02-01 14:06:36,816 [ceph1][DEBUG ]     "modified": "0.000000",
2014-02-01 14:06:36,816 [ceph1][DEBUG ]     "mons": [
2014-02-01 14:06:36,817 [ceph1][DEBUG ]       {
2014-02-01 14:06:36,818 [ceph1][DEBUG ]         "addr":
"192.168.111.11:6789/0",
2014-02-01 14:06:36,818 [ceph1][DEBUG ]         "name": "ceph1",
2014-02-01 14:06:36,819 [ceph1][DEBUG ]         "rank": 0
2014-02-01 14:06:36,820 [ceph1][DEBUG ]       }
2014-02-01 14:06:36,821 [ceph1][DEBUG ]     ]
2014-02-01 14:06:36,822 [ceph1][DEBUG ]   },
2014-02-01 14:06:36,826 [ceph1][DEBUG ]   "name": "ceph1",
2014-02-01 14:06:36,826 [ceph1][DEBUG ]   "outside_quorum": [],
2014-02-01 14:06:36,826 [ceph1][DEBUG ]   "quorum": [
2014-02-01 14:06:36,827 [ceph1][DEBUG ]     0
2014-02-01 14:06:36,827 [ceph1][DEBUG ]   ],
2014-02-01 14:06:36,827 [ceph1][DEBUG ]   "rank": 0,
2014-02-01 14:06:36,827 [ceph1][DEBUG ]   "state": "leader",
2014-02-01 14:06:36,828 [ceph1][DEBUG ]   "sync_provider": []
2014-02-01 14:06:36,828 [ceph1][DEBUG ] }
2014-02-01 14:06:36,828 [ceph1][DEBUG ]

********************************************************************************
2014-02-01 14:06:36,829 [ceph1][INFO  ] monitor: mon.ceph1 is running
2014-02-01 14:06:36,830 [ceph1][INFO  ] Running command: sudo ceph
--cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph1.asok
mon_status
2014-02-01 14:06:37,005 [ceph_deploy.mon][INFO  ] processing monitor
mon.ceph1
2014-02-01 14:06:37,079 [ceph1][DEBUG ] connected to host: ceph1
2014-02-01 14:06:37,081 [ceph1][INFO  ] Running command: sudo ceph
--cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.ceph1.asok
mon_status
2014-02-01 14:06:37,258 [ceph_deploy.mon][INFO  ] mon.ceph1 monitor has
reached quorum!
2014-02-01 14:06:37,259 [ceph_deploy.mon][INFO  ] all initial monitors
are
running and have formed quorum
2014-02-01 14:06:37,266 [ceph_deploy.mon][INFO  ] Running gatherkeys...
2014-02-01 14:06:37,268 [ceph_deploy.gatherkeys][DEBUG ] Checking ceph1
for
/etc/ceph/ceph.client.admin.keyring
2014-02-01 14:06:37,336 [ceph1][DEBUG ] connected to host: ceph1
2014-02-01 14:06:37,340 [ceph1][DEBUG ] detect platform information from
remote host
2014-02-01 14:06:37,373 [ceph1][DEBUG ] detect machine type
2014-02-01 14:06:37,383 [ceph1][DEBUG ] fetch remote file

2014-02-01 14:06:37,385 [ceph_deploy.gatherkeys][WARNING] Unable to find
/etc/ceph/ceph.client.admin.keyring on ['ceph1']
2014-02-01 14:06:37,391 [ceph_deploy.gatherkeys][DEBUG ] Have
ceph.mon.keyring
2014-02-01 14:06:37,398 [ceph_deploy.gatherkeys][DEBUG ] Checking ceph1
for
/var/lib/ceph/bootstrap-osd/ceph.keyring
2014-02-01 14:06:37,468 [ceph1][DEBUG ] connected to host: ceph1
2014-02-01 14:06:37,471 [ceph1][DEBUG ] detect platform information from
remote host
2014-02-01 14:06:37,506 [ceph1][DEBUG ] detect machine type
2014-02-01 14:06:37,514 [ceph1][DEBUG ] fetch remote file

2014-02-01 14:06:37,516 [ceph_deploy.gatherkeys][WARNING] Unable to find
/var/lib/ceph/bootstrap-osd/ceph.keyring on ['ceph1']
2014-02-01 14:06:37,523 [ceph_deploy.gatherkeys][DEBUG ] Checking ceph1
for
/var/lib/ceph/bootstrap-mds/ceph.keyring
2014-02-01 14:06:37,591 [ceph1][DEBUG ] connected to host: ceph1
2014-02-01 14:06:37,594 [ceph1][DEBUG ] detect platform information from
remote host
2014-02-01 14:06:37,627 [ceph1][DEBUG ] detect machine type
2014-02-01 14:06:37,636 [ceph1][DEBUG ] fetch remote file

2014-02-01 14:06:37,639 [ceph_deploy.gatherkeys][WARNING] Unable to find
/var/lib/ceph/bootstrap-mds/ceph.keyring on ['ceph1']

------------------------------------------------------------------------------------
Does it end here? seems like the output was trimmed...











(2014/02/03 22:26), Alfredo Deza wrote:
On Sun, Feb 2, 2014 at 12:18 AM, Kei.masumoto <kei.masumoto@xxxxxxxxx>
wrote:
Hi,

I am newbie of ceph, now I am trying to deploy following
"http://ceph.com/docs/master/start/quick-ceph-deploy/";
ceph1, ceph2 and ceph3 exists according to the above tutorial. I got a
WARNING message when I exec ceph-deploy "mon create-initial".

2014-02-01 14:06:37,385 [ceph_deploy.gatherkeys][WARNING] Unable to
find
/etc/ceph/ceph.client.admin.keyring on ['ceph1']
2014-02-01 14:06:37,516 [ceph_deploy.gatherkeys][WARNING] Unable to
find
/var/lib/ceph/bootstrap-osd/ceph.keyring on ['ceph1']
2014-02-01 14:06:37,639 [ceph_deploy.gatherkeys][WARNING] Unable to
find
/var/lib/ceph/bootstrap-mds/ceph.keyring on ['ceph1']

Thinking about when those 3 keyrings should be created, I thins
"ceph-deploy mon create " is a right timing for keyring creation. I
checked my environment, and found
/etc/ceph/ceph.client.admin.keyring.14081.tmp. It looks like this file
is created by ceph-create-keys on executing stop ceph-all && start
ceph-all. but ceph-create-keys never finishes.
ceph-deploy tries to help here a lot with create-initial, and although
the warnings are useful,
they are only good depending on the context of the rest of the output.

When the whole process completes, does ceph-deploy say all mons are up
and running?

It would be better to paste the complete output of the call so we can
see the details.
When I execute ceph-create-keys manually, it continues to generate
below
log, looks like waiting reply...

2014-02-01 20:13:02.847737 7f55e81a4700  0 -- :/1001774 >>
192.168.11.8:6789/0 pipe(0x7f55e4024400 sd=3 :0 s=1 pgs=0 cs=0 l=1
c=0x7f55e4024660).fault

Since I found that mon listens 6789, so I strace mon, then mon also
waiting something...

root@ceph1:~/src/ceph-0.56.7# strace -p 1047
Process 1047 attached - interrupt to quit
futex(0x7f37c14839d0, FUTEX_WAIT, 1102, NULL

I have no idea what situation should be, any hints?

P.S. somebody give me an adivce to check below, but I dont see any from
here.
root@ceph1:~/my-cluster# ceph daemon mon.`hostname` mon_status
{ "name": "ceph1",
     "rank": 0,
     "state": "leader",
     "election_epoch": 1,
     "quorum": [
           0],
     "outside_quorum": [],
     "extra_probe_peers": [],
     "sync_provider": [],
     "monmap": { "epoch": 1,
         "fsid": "26835656-6b29-455d-9d1f-545cad8f1e23",
         "modified": "0.000000",
         "created": "0.000000",
         "mons": [
               { "rank": 0,
                 "name": "ceph1",
                 "addr": "192.168.111.11:6789\/0"}]}}


Kei
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux